Pascal Pfeiffer, Principal Data Scientist, H2O.ai
H2O Open Source GenAI World SF 2023
This talk dives into the expansive ecosystem of Large Language Models (LLMs), offering practitioners an insightful guide to various relevant applications, from natural language understanding to creative content generation. While exploring use cases across different industries, it also honestly addresses the current limitations of LLMs and anticipates future advancements.
This document provides information about a bootcamp to build applications using Large Language Models (LLMs). The bootcamp consists of 11 modules covering topics such as introduction to generative AI, text analytics techniques, neural network models for natural language processing, transformer models, embedding retrieval, semantic search, prompt engineering, fine-tuning LLMs, orchestration frameworks, the LangChain application platform, and a final project to build a custom LLM application. The bootcamp will be held in various locations and dates between September 2023 and January 2024.
UNLEASHING INNOVATION Exploring Generative AI in the Enterprise.pdfHermes Romero
The document provides an overview of generative AI, including its key concepts and applications. It discusses transformer models versus neural networks, explaining that transformer models use self-attention to capture long-range dependencies in sequential data like text. Large language models (LLMs) based on the transformer architecture have shown strong performance in natural language generation tasks. The document outlines the evolution of generative AI techniques from early machine learning to modern large pretrained models. It also surveys some commercial generative AI applications in industries like healthcare, finance, and gaming.
Challenges in AI LLMs adoption in the EnterpriseGeorge Bara
The presentation "ITDays_2023_GeorgeBara" discusses challenges in adopting AI large language models (LLMs) in enterprise settings.
The presentation covers:
1. **Challenges in AI LLMs adoption**: It highlights the noise in the current AI landscape and questions the practical use of AI in real businesses.
2. **The DNA of an Enterprise**: Defines enterprise sizes and discusses the new solutions adoption process, emphasizing effective integration and minimizing disruption.
3. **Enterprise-Grade**: Lists qualities like robustness, reliability, scalability, performance, security, and support that are essential for enterprise-grade solutions.
4. **What are LLMs?**: Describes the pre-ChatGPT era with BERT, a model used for language understanding, and details its enterprise applications.
5. **LLM use-cases before ChatGPT**: Focuses on data triage, process automation, knowledge management, and the augmentation of business operations.
6. **EU Digital Decade Report**: Points out that AI adoption in Europe is slow and might not meet the 2030 targets.
7. **Adoption Challenges**: Addresses top challenges such as data security, predictability, performance, control, regulatory compliance, ethics, sustainability, and ROI.
8. **Conclusion**: Reflects on the slow adoption of AI in enterprises, suggesting that a surge might occur once the technology matures and is ready for enterprise use.
The presenter concludes by stating that despite the hype around technologies like ChatGPT, enterprises are cautious and will adopt new technologies at their own pace. He anticipates a gradual then sudden adoption pattern once LLMs are proven to be enterprise-ready.
How Does Generative AI Actually Work? (a quick semi-technical introduction to...ssuser4edc93
This document provides a technical introduction to large language models (LLMs). It explains that LLMs are based on simple probabilities derived from their massive training corpora, containing trillions of examples. The document then discusses several key aspects of how LLMs work, including that they function as a form of "lossy text compression" by encoding patterns and relationships in their training data. It also outlines some of the key elements in the architecture and training of the most advanced LLMs, such as GPT-4, focusing on their huge scale, transformer architecture, and use of reinforcement learning from human feedback.
Active Retrieval Augmented Generation.pdfPo-Chuan Chen
FLARE proposes a method called Forward-Looking Active REtrieval augmented generation (FLARE) that iteratively retrieves information during text generation based on the predicted upcoming sentence. FLARE uses the predicted next sentence as a query to retrieve documents if it contains low-confidence tokens, then regenerates the sentence. Experiments show FLARE outperforms baselines on multiple knowledge-intensive tasks. However, FLARE did not significantly improve performance on a short-text dataset where continual retrieval of disparate information may not be needed.
Generative AI Masterclass - Model Risk Management.pptxSri Ambati
Here are some key points about benchmarking and evaluating generative AI models like large language models:
- Foundation models require large, diverse datasets to be trained on in order to learn broad language skills and knowledge. Fine-tuning can then improve performance on specific tasks.
- Popular benchmarks evaluate models on tasks involving things like commonsense reasoning, mathematics, science questions, generating truthful vs false responses, and more. This helps identify model capabilities and limitations.
- Custom benchmarks can also be designed using tools like Eval Studio to systematically test models on specific applications or scenarios. Both automated and human evaluations are important.
- Leaderboards like HELM aggregate benchmark results to compare how different models perform across a wide range of tests and metrics.
This presentation presents an overview of the challenges and opportunities of generative artificial intelligence in Web3. It includes a brief research history of generative AI as well as some of its immediate applications in Web3.
Retrieval Augmented Generation in Practice: Scalable GenAI platforms with k8s...Mihai Criveti
Mihai is the Principal Architect for Platform Engineering and Technology Solutions at IBM, responsible for Cloud Native and AI Solutions. He is a Red Hat Certified Architect, CKA/CKS, a leader in the IBM Open Innovation community, and advocate for open source development. Mihai is driving the development of Retrieval Augmentation Generation platforms, and solutions for Generative AI at IBM that leverage WatsonX, Vector databases, LangChain, HuggingFace and open source AI models.
Mihai will share lessons learned building Retrieval Augmented Generation, or “Chat with Documents” platforms and APIs that scale, and deploy on Kubernetes. His talk will cover use cases for Generative AI, limitations of Large Language Models, use of RAG, Vector Databases and Fine Tuning to overcome model limitations and build solutions that connect to your data and provide content grounding, limit hallucinations and form the basis of explainable AI. In terms of technology, he will cover LLAMA2, HuggingFace TGIS, SentenceTransformers embedding models using Python, LangChain, and Weaviate and ChromaDB vector databases. He’ll also share tips on writing code using LLM, including building an agent for Ansible and containers.
Scaling factors for Large Language Model Architectures:
• Vector Database: consider sharding and High Availability
• Fine Tuning: collecting data to be used for fine tuning
• Governance and Model Benchmarking: how are you testing your model performance
over time, with different prompts, one-shot, and various parameters
• Chain of Reasoning and Agents
• Caching embeddings and responses
• Personalization and Conversational Memory Database
• Streaming Responses and optimizing performance. A fine tuned 13B model may
perform better than a poor 70B one!
• Calling 3rd party functions or APIs for reasoning or other type of data (ex: LLMs are
terrible at reasoning and prediction, consider calling other models)
• Fallback techniques: fallback to a different model, or default answers
• API scaling techniques, rate limiting, etc.
• Async, streaming and parallelization, multiprocessing, GPU acceleration (including
embeddings), generating your API using OpenAPI, etc.
Chat GPT 4 can pass the American state bar exam, but before you go expecting to see robot lawyers taking over the courtroom, hold your horses cowboys – we're not quite there yet. That being said, AI is becoming increasingly more human-like, and as a VC we need to start thinking about how this new wave of technology is going to affect the way we build and run businesses. What do we need to do differently? How can we make sure that our investment strategies are reflecting these changes? It's a brave new world out there, and we’ve got to keep the big picture in mind!
Sharing here with you what we at Cavalry Ventures found out during our Generative AI deep dive.
Unlocking the Power of Generative AI An Executive's Guide.pdfPremNaraindas1
Generative AI is here, and it can revolutionize your business. With its powerful capabilities, this technology can help companies create more efficient processes, unlock new insights from data, and drive innovation. But how do you make the most of these opportunities?
This guide will provide you with the information and resources needed to understand the ins and outs of Generative AI, so you can make informed decisions and capitalize on the potential. It covers important topics such as strategies for leveraging large language models, optimizing MLOps processes, and best practices for building with Generative AI.
This session was presented at the AWS Community Day in Munich (September 2023). It's for builders that heard the buzz about Generative AI but can’t quite grok it yet. Useful if you are eager to connect the dots on the Generative AI terminology and get a fast start for you to explore further and navigate the space. This session is largely product agnostic and meant to give you the fundamentals to get started.
Generative AI: Past, Present, and Future – A Practitioner's PerspectiveHuahai Yang
Generative AI: Past, Present, and Future – A Practitioner's Perspective
As the academic realm grapples with the profound implications of generative AI
and related applications like ChatGPT, I will present a grounded view from my
experience as a practitioner. Starting with the origins of neural networks in
the fields of logic, psychology, and computer science, I trace its history and
align it within the wider context of the pursuit of artificial intelligence.
This perspective will also draw parallels with historical developments in
psychology. Against this backdrop, I chart a proposed trajectory for the future.
Finally, I provide actionable insights for both academics and enterprising
individuals in the field.
LLMOps for Your Data: Best Practices to Ensure Safety, Quality, and CostAggregage
Join Shreya Rajpal, CEO of Guardrails AI, and Travis Addair, CTO of Predibase, in this exclusive webinar to learn all about leveraging the part of AI that constitutes your IP – your data – to build a defensible AI strategy for the future!
generative-ai-fundamentals and Large language modelsAdventureWorld5
Thank you for the detailed review of the protein bars. I'm glad to hear you and your family are enjoying them as a healthy snack and meal replacement option. A couple suggestions based on your feedback:
- For future orders, you may want to check the expiration dates to help avoid any dried out bars towards the end of the box. Freshness is key to maintaining the moist texture.
- When introducing someone new to the bars, selecting one in-person if possible allows checking the flexibility as an indicator it's moist inside. This could help avoid a disappointing first impression from a dry sample.
- Storing opened boxes in an airtight container in the fridge may help extend the freshness even further when you can't
Delve into this insightful article to explore the current state of generative AI, its ethical implications, and the power of generative AI models across various industries.
Generative AI is evolving rapidly and disrupting marketing and sales in several ways:
1) It can leverage large datasets to identify new audience segments and automatically generate personalized outreach content at scale.
2) Within the sales process, it provides continuous support through tasks like hyper-personalized messaging, virtual assistance, and predictive insights.
3) It also has applications in customer onboarding, retention, and success analytics through tools like dynamic content and customer journey mapping.
Commercial leaders anticipate moderate to significant impact from generative AI use cases and most expect to utilize such solutions extensively in the next two years. Effective companies are prioritizing technologies like generative AI to improve performance.
This document discusses AI and ChatGPT. It begins with an introduction to David Cieslak and his company RKL eSolutions, which provides ERP sales and consulting. It then provides definitions for key AI concepts like artificial intelligence, generative AI, large language models, and ChatGPT. The document discusses OpenAI's ChatGPT tool and how it works. It covers prompts, commands, and potential uses and impacts of generative AI technologies. Finally, it discusses concerns regarding generative AI and the future of life institute's call for more oversight of advanced AI.
Here is a draft email:
Subject: Automate key processes in automotive manufacturing with UiPath
Dear Tom,
My name is Ed Challis from UiPath. I understand from our mutual connection that you are the Automation Program Manager at BMW, focusing on implementing robotic process automation (RPA).
I wanted to share how some of our automotive manufacturing customers are leveraging UiPath to drive efficiencies in their operations. Specifically:
Quality inspection automation: One customer automated visual inspections on the production line to reduce defects and speed up issue resolution. This helped improve quality standards.
Supply chain management: Another customer automated PO matching, invoice processing and inventory management across their suppliers globally. This
BDW17 London - Abed Ajraou - First Utility - Putting Data Science in your Bus...Big Data Week
Data Science is now well established in our businesses, and everyone considers data as a key asset and critical for our competitiveness.
However, Data Science is not easy to manage, very often projects failed and the investment made is not seeing as profitable.
The aim of this talk is to share the knowledge in different areas:
* avoid classical mistakes in Data Science
* use the right Big Data technology
* apply the right methodology
* make the Data Science team more efficient
DataScientist Job : Between Myths and Reality.pdfJedha Bootcamp
Swipe through the smoke and mirrors and learn about the "sexiest job of the 21st century" with Nicola, Machine Learning Scientist @ Bumble
✨ Artificial Intelligence? Business Intelligence? Data Science? What do these terms sound like when put into action at one of the world's most forefront dating platforms? Jedha is proud to host an evening with Nicola Ghio, Senior Machine Learning Scientist at Bumble, who will give us a "peek behind the curtain" into what this enviable job title looks like in practice.
😎 Nicola will share some of his experiences working at Bumble. 🎯 Hear first-hand about Bumble's harassment and toxic imaging detector as well as the real skills required to work in the industry. We also look forward to hearing about Nicola's personal story, his background and his advice for those that want to dive deeper into the world of tech.
Meet Jedha 😍 Your Data and Cyber Security Bootcamp, ranked #1 in Europe (Switch Up). Our mission is to demystify the world of tech and to make its skills accessible to anyone who desires to learn. We have courses suited to all ambitions and skill levels: From beginners who have never typed a line of code in their lives right through to skilled tech professionals who want to achieve mastery. Our methods and teachers help to unlock human potential in the unlimited world of tech.
This is a class presentation on the Emerging Technology Analysis Canvas (ETAC), a framework to assess emerging technologies. It was created for students in a course on emerging technologies.
Any truly important technology creates fear and uncertainty. By this measure AI is going to become truly significant. However to achieve this we must close the cognitive concept gap.
Salesforce Architect Group, Frederick, United States July 2023 - Generative A...NadinaLisbon1
Joined our community-led event to dive into the world of Artificial Intelligence (AI)! Whether you were just starting your AI journey or already familiar with its concepts, one thing was certain: AI was reshaping the future of work. This enablement session was your chance to level up your skills and stay ahead in that rapidly evolving landscape.
As AI news continues to dominate headlines, it's natural to have questions and concerns about its impact on our lives. Will AI take over human jobs? Will it render us obsolete? Rest assured, the outlook is far brighter than you may think. Rather than replacing humans, AI is designed to enhance our capabilities and work alongside us. It won't be replacing marketers, service representatives, or salespeople—it will be empowering them to achieve even greater results. Companies across industries recognize this potential and are embracing AI to unlock new levels of performance.
During this enablement session, you'll have the opportunity to explore how AI advancements can positively influence your professional journey and daily life. We'll debunk common misconceptions, address fears, and showcase real-world examples of how successful AI implementation leads to workforce augmentation rather than replacement. Be prepared to gain valuable insights and practical knowledge that will help you navigate the AI landscape with confidence.
MongoDB World 2018: Building Intelligent Apps with MongoDB & Google CloudMongoDB
Building intelligent apps involves combining real-time analytics, machine learning, and artificial intelligence to provide personalized recommendations and automate tasks for customers. Developers can use MongoDB and Google Cloud to build intelligent apps in 3 steps: 1) create a base ecommerce app, 2) add a recommendation engine using machine learning, and 3) enable shopping via chat with artificial intelligence. This brings data scientists and developers together to create applications that understand and assist customers.
[DSC Europe 22] Avoid mistakes building AI products - Karol PrzystalskiDataScienceConferenc1
Based on Gartner's research, 85% of AI projects fail. In this talk, we show the most common mistakes made by the managers, developers, and data scientists while building AI products. We go through ten case studies of products that failed and analyze the reasons for each failure. We also present how to avoid such mistakes and deliver a successful AI product by introducing a few lifecycle changes.
Machine Learning has become a must to improve insight, quality and time to market. But it's also been called the 'high interest credit card of technical debt' with challenges in managing both how it's applied and how its results are consumed.
The document announces an AI and OpenPOWER meetup to take place on March 25th, 2018 from 4-7:30pm at the h2o.AI headquarters in Mountain View, CA. The meetup will feature prominent speakers from industry, research, and the financial sector who will discuss advances in deep learning tools and techniques. Key speakers include Ganesan Narayanasamy from IBM who will discuss OpenPOWER activities and supercomputers, and Sudha Jamthe from IoTDisruptions.com who will discuss AI trends towards a driverless world.
This document provides guidance on building a career in AI through three key steps: learning foundational skills, working on projects, and finding a job. It discusses each step in detail with chapters focused on learning technical skills, scoping AI projects, and using projects to complement career goals. The overall message is that an AI career requires lifelong learning, gaining experience through meaningful projects, and navigating an evolving job market. Building a supportive community is also important for support throughout the career journey.
*Uses of AI and data science can be found in almost any situation that produces data
* More uses for custom AI applications and data-derived
insights than for traditional software engineering
* Literacy in AI-oriented coding will be more valuable than traditional coding
Machine learning is a term thrown around in technology circles with an ever-increasing intensity. Major
technology companies have attached themselves to
this buzzword to receive capital investments, and every
major technology company is pushing its even shinier
parentartificial intelligence (AI).
My programming and machine learning linked in notes 2021 part 1Vedran Markulj
The document discusses machine learning platforms and tools. It notes that while AutoML promises code-free machine learning, it only solves the easiest part and data scientists still need to understand data science and have functioning pipelines. Good machine learning work involves engineering problems like data analysis and feature engineering, not just the model work. There is no standard machine learning platform yet due to the field still being early, so companies need to formulate realistic plans and not expect best practices to be defined. Heuristics can get companies started without machine learning if they lack data.
The document discusses anti-patterns and worst practices in software development. Some examples covered include static cling pattern, flags over objects, premature optimization, copy-paste-compile, and reinventing the wheel. It also shares lessons learned from experiences, such as being mindful of date times across time zones, avoiding building SQL from untrusted inputs, and not being too cute with test data. Overall, the document aims to help developers learn from the mistakes of others and adopt better practices.
SDD2017 - 03 Abed Ajraou - putting data science in your business a first uti...Dario Mangano
This document discusses putting data science solutions into business practices. It emphasizes the importance of starting with a clear business problem rather than just focusing on the data. It also recommends adopting the right technology, mindset, and methodology. For methodology, it advocates an iterative approach using techniques like exploratory analysis, feature engineering, and machine learning algorithms like gradient boosting. It also discusses automating machine learning tasks and gaining efficiency through collaborative data science platforms.
The document discusses the Lean Startup methodology for building startups with a focus on minimizing risk through continuous experimentation and customer feedback. It contrasts two approaches - one that failed after 5 years and $40M by making assumptions without validating them, and one that succeeded by rapidly iterating and testing hypotheses with customers. The Lean Startup process emphasizes small batches, continuous deployment, A/B testing, and the "Five Whys" technique to drive learning and improve faster than traditional approaches.
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo DaySri Ambati
This document provides an overview of H2O.ai, an AI company that offers products and services to democratize AI. It mentions that H2O products are backed by 10% of the world's top data scientists from Kaggle and that H2O has customers in 7 of the top 10 banks, 4 of the top 10 insurance companies, and top manufacturing companies. It also provides details on H2O's founders, funding, customers, products, and vision to make AI accessible to more organizations.
LLMOps: Match report from the top of the 5thSri Ambati
The document discusses LLMOps (Large Language Model Operations) compared to traditional MLOps. Some key points:
- LLMOps and MLOps face similar challenges across the development lifecycle, but LLMOps requires more GPU resources and integration is faster due to more models in each application. Evaluation is also less clear.
- The LLMOps field is around the 5th generation of models, with debates around proprietary vs open source models, and balancing privacy, cost and control.
- LLMOps platforms are emerging to provide solutions for tasks like prompting, embedding databases, evaluation, and governance, similar to how MLOps platforms have evolved.
Building, Evaluating, and Optimizing your RAG App for ProductionSri Ambati
The document discusses optimizing question answering systems called RAG (Retrieve-and-Generate) stacks. It outlines challenges with naive RAG approaches and proposes solutions like improved data representations, advanced retrieval techniques, and fine-tuning large language models. Table stakes optimizations include tuning chunk sizes, prompt engineering, and customizing LLMs. More advanced techniques involve small-to-big retrieval, multi-document agents, embedding fine-tuning, and LLM fine-tuning.
Building LLM Solutions using Open Source and Closed Source Solutions in Coher...Sri Ambati
Sandeep Singh, Head of Applied AI Computer Vision, Beans.ai
H2O Open Source GenAI World SF 2023
In the modern era of machine learning, leveraging both open-source and closed-source solutions has become paramount for achieving cutting-edge results. This talk delves into the intricacies of seamlessly integrating open-source Large Language Model (LLM) solutions like Vicuna, Falcon, and Llama with industry giants such as ChatGPT and Google's Palm. As the demand for fine-tuned and specialized datasets grows, it is imperative to understand the synergy between these tools. Attendees will gain insights into best practices for building and enriching datasets tailored for fine-tuning tasks, ensuring that their LLM projects are both robust and efficient. Through real-world examples and hands-on demonstrations, this talk will equip attendees with the knowledge to harness the power of both open and closed-source tools in a coherent and effective manner.
Patrick Hall, Professor, AI Risk Management, The George Washington University
H2O Open Source GenAI World SF 2023
Language models are incredible engineering breakthroughs but require auditing and risk management before productization. These systems raise concerns about toxicity, transparency and reproducibility, intellectual property licensing and ownership, disinformation and misinformation, supply chains, and more. How can your organization leverage these new tools without taking on undue or unknown risks? While language models and associated risk management are in their infancy, a small number of best practices in governance and risk are starting to emerge. If you have a language model use case in mind, want to understand your risks, and do something about them, this presentation is for you!
Dr. Alexy Khrabrov, Open Source Science Community Director, IBM
H2O Open Source GenAI World SF 2023
In this talk, Dr. Alexy Khrabrov, recently elected Chair of the new Generative AI Commons at Linux Foundation for AI & Data, outlines the OSS AI landscape, challenges, and opportunities. With new models and frameworks being unveiled weekly, one thing remains constant: community building and validation of all aspects of AI is key to reliable and responsible AI we can use for business and society needs. Industrial AI is one key area where such community validation can prove invaluable.
The document announces the launch of the H2O GenAI App Store, which provides a collection of applications that make it easier for average users to leverage large language models through custom interfaces for specific tasks like getting gardening advice or feedback on code. The app store is designed to accelerate the development of these GenAI apps using the H2O Wave platform and provides access to H2OGPTE for retrieval augmented generation and language model calls. Developers can also contribute their own apps through the GitHub repository listed.
Applied Gen AI for the Finance Vertical Sri Ambati
Megan Kurka, Vice President, Customer Data Scientist, H2O.ai
H2O Open Source GenAI World SF 2023
Discover the transformative power of Applied Gen AI. Learn how the H2O team builds customized applications and workflows that integrate capabilities of Gen AI and AutoML specifically designed to address and enhance financial use cases. Explore real world examples, learn best practices, and witness firsthand how our innovative solutions are reshaping the landscape of finance technology.
This document discusses techniques for improving language models (LLMs) discussed in recent papers. It describes building blocks of LLMs like fine-tuning, foundation training, memory, and databases. Specific techniques covered include LIMA which uses 1,000 carefully curated examples, instruction backtranslation to generate question-answer pairs, fine-tuning models on API examples like Gorilla, and reducing false answers through techniques like not agreeing with incorrect user opinions. The goal is to discuss cutting edge tricks to build better LLMs.
KGM Mastering Classification and Regression with LLMs: Insights from Kaggle C...Sri Ambati
This document discusses using large language models (LLMs) for text classification tasks. It begins by describing how LLMs are commonly used for text generation and question answering. For classification, models are usually trained supervised on labeled data. The document then explores using LLMs for zero-shot classification without training, and techniques like fine-tuning LLMs on tasks to improve performance. It provides an example of fine-tuning an LLM on a financial sentiment dataset. The document concludes by describing H2O.ai's LLM Studio tool for fine-tuning and a few Kaggle competitions where LLMs achieved success in text classification.
1) Generative AI (GenAI) enables the creation of novel content by learning patterns in unstructured data rather than labeling outputs like traditional AI.
2) Both traditional and generative AI models lack transparency and may contain biases, but generative models can additionally hallucinate or leak private information.
3) To interpret generative models, researchers evaluate accuracy globally by checking for hallucinations or undesirable content, and locally by confirming the quality of individual responses.
Introducción al Aprendizaje Automatico con H2O-3 (1)Sri Ambati
En esta reunión virtual, damos una introducción a la plataforma de aprendizaje automático de código abierto número 1, H2O-3 y te mostramos cómo puedes usarla para desarrollar modelos para resolver diferentes casos de uso.
From Rapid Prototypes to an end-to-end Model Deployment: an AI Hedge Fund Use...Sri Ambati
Numerai is an open, crowd-sourced hedge fund powered by predictions from data scientists around the world. In return, participants are rewarded with weekly payouts in crypto.
In this talk, Joe will give an overview of the Numerai tournament based on his own experience. He will then explain how he automates the time-consuming tasks such as testing different modelling strategies, scoring new datasets, submitting predictions to Numerai as well as monitoring model performance with H2O Driverless AI and R.
AI Foundations Course Module 1 - Shifting to the Next Step in Your AI Transfo...Sri Ambati
In this session, you will learn about what you should do after you’ve taken an AI transformation baseline. Over the span of this session, we will discuss the next steps in moving toward AI readiness through alignment of talent and tools to drive successful adoption and continuous use within an organization.
To find additional videos on AI courses, earn badges, join the courses at H2O.ai Learning Center: https://training.h2o.ai/products/ai-foundations-course
To find the Youtube video about this presentation: https://youtu.be/K1Cl3x3rd8g
Speaker:
Chemere Davis (H2O.ai - Senior Data Scientist Training Specialist)
AI Foundations Course Module 1 - An AI Transformation JourneySri Ambati
The chances of successfully implementing AI strategies within an organization significantly improve when you can recognize where your organization is on the maturity scale. Over this course, you will learn the keys to unlocking value with AI which include asking the right questions about the problems you are solving and ensuring you have the right cross-section of talent, tools, and resources. By the end of this module, you should be able to recognize where your organization is on the AI transformation spectrum and identify some strategies that can get you to the next stage in your journey.
To find additional videos on AI courses, earn badges, join the courses at H2O.ai Learning Center: https://training.h2o.ai/products/ai-foundations-course
To find the Youtube video about this presentation: https://youtu.be/PJgr2epM6qs
Speakers:
Chemere Davis (H2O.ai - Senior Data Scientist Training Specialist)
Ingrid Burton (H2O.ai - CMO)
ML Model Deployment and Scoring on the Edge with Automatic ML & DFSri Ambati
Machine Learning Model Deployment and Scoring on the Edge with Automatic Machine Learning and Data Flow
YouTube Video URL: https://youtu.be/gB0bTH-L6DE
Deploying Machine Learning models to the edge can present significant ML/IoT challenges centered around the need for low latency and accurate scoring on minimal resource environments. H2O.ai's Driverless AI AutoML and Cloudera Data Flow work nicely together to solve this challenge. Driverless AI automates the building of accurate Machine Learning models, which are deployed as light footprint and low latency Java or C++ artifacts, also known as a MOJO (Model Optimized). And Cloudera Data Flow leverage Apache NiFi that offers an innovative data flow framework to host MOJOs to make predictions on data moving on the edge.
Scaling & Managing Production Deployments with H2O ModelOpsSri Ambati
This presentation was made on June 30th, 2020.
Recording of the presentation is available here: https://youtu.be/9LajqAL_CU8
As enterprises “make their own AI”, a new set of challenges emerge. Maintaining reproducibility, traceability, and verifiability of machine learning models, as well as recording experiments, tracking insights, and reproducing results, are key. Collaboration between teams is also necessary as “model factories” are created for enterprise-wide model data science efforts. Additionally, monitoring of models ensures that drift or performance degradation is addressed with either retraining or model updates. Finally, data and model lineage in case of rollbacks or addressing regulatory compliance is necessary.
H2O ModelOps delivers centralized catalog and management, deployment, monitoring, collaboration, and administration of machine learning models. In this webinar, we learn how H2O can assist with operationalizing, scaling and managing production deployments.
Speaker's Bio:
Felix is a part of the Customer Success team in Asia Pacific at H2O.ai. An engineer and an IIM alumni, Felix has held prominent positions in the data science industry.
Choose our Linux Web Hosting for a seamless and successful online presencerajancomputerfbd
Our Linux Web Hosting plans offer unbeatable performance, security, and scalability, ensuring your website runs smoothly and efficiently.
Visit- https://onliveserver.com/linux-web-hosting/
Coordinate Systems in FME 101 - Webinar SlidesSafe Software
If you’ve ever had to analyze a map or GPS data, chances are you’ve encountered and even worked with coordinate systems. As historical data continually updates through GPS, understanding coordinate systems is increasingly crucial. However, not everyone knows why they exist or how to effectively use them for data-driven insights.
During this webinar, you’ll learn exactly what coordinate systems are and how you can use FME to maintain and transform your data’s coordinate systems in an easy-to-digest way, accurately representing the geographical space that it exists within. During this webinar, you will have the chance to:
- Enhance Your Understanding: Gain a clear overview of what coordinate systems are and their value
- Learn Practical Applications: Why we need datams and projections, plus units between coordinate systems
- Maximize with FME: Understand how FME handles coordinate systems, including a brief summary of the 3 main reprojectors
- Custom Coordinate Systems: Learn how to work with FME and coordinate systems beyond what is natively supported
- Look Ahead: Gain insights into where FME is headed with coordinate systems in the future
Don’t miss the opportunity to improve the value you receive from your coordinate system data, ultimately allowing you to streamline your data analysis and maximize your time. See you there!
The DealBook is our annual overview of the Ukrainian tech investment industry. This edition comprehensively covers the full year 2023 and the first deals of 2024.
Fluttercon 2024: Showing that you care about security - OpenSSF Scorecards fo...Chris Swan
Have you noticed the OpenSSF Scorecard badges on the official Dart and Flutter repos? It's Google's way of showing that they care about security. Practices such as pinning dependencies, branch protection, required reviews, continuous integration tests etc. are measured to provide a score and accompanying badge.
You can do the same for your projects, and this presentation will show you how, with an emphasis on the unique challenges that come up when working with Dart and Flutter.
The session will provide a walkthrough of the steps involved in securing a first repository, and then what it takes to repeat that process across an organization with multiple repos. It will also look at the ongoing maintenance involved once scorecards have been implemented, and how aspects of that maintenance can be better automated to minimize toil.
論文紹介:A Systematic Survey of Prompt Engineering on Vision-Language Foundation ...Toru Tamaki
Jindong Gu, Zhen Han, Shuo Chen, Ahmad Beirami, Bailan He, Gengyuan Zhang, Ruotong Liao, Yao Qin, Volker Tresp, Philip Torr "A Systematic Survey of Prompt Engineering on Vision-Language Foundation Models" arXiv2023
https://arxiv.org/abs/2307.12980
The Rise of Supernetwork Data Intensive ComputingLarry Smarr
Invited Remote Lecture to SC21
The International Conference for High Performance Computing, Networking, Storage, and Analysis
St. Louis, Missouri
November 18, 2021
UiPath Community Day Kraków: Devs4Devs ConferenceUiPathCommunity
We are honored to launch and host this event for our UiPath Polish Community, with the help of our partners - Proservartner!
We certainly hope we have managed to spike your interest in the subjects to be presented and the incredible networking opportunities at hand, too!
Check out our proposed agenda below 👇👇
08:30 ☕ Welcome coffee (30')
09:00 Opening note/ Intro to UiPath Community (10')
Cristina Vidu, Global Manager, Marketing Community @UiPath
Dawid Kot, Digital Transformation Lead @Proservartner
09:10 Cloud migration - Proservartner & DOVISTA case study (30')
Marcin Drozdowski, Automation CoE Manager @DOVISTA
Pawel Kamiński, RPA developer @DOVISTA
Mikolaj Zielinski, UiPath MVP, Senior Solutions Engineer @Proservartner
09:40 From bottlenecks to breakthroughs: Citizen Development in action (25')
Pawel Poplawski, Director, Improvement and Automation @McCormick & Company
Michał Cieślak, Senior Manager, Automation Programs @McCormick & Company
10:05 Next-level bots: API integration in UiPath Studio (30')
Mikolaj Zielinski, UiPath MVP, Senior Solutions Engineer @Proservartner
10:35 ☕ Coffee Break (15')
10:50 Document Understanding with my RPA Companion (45')
Ewa Gruszka, Enterprise Sales Specialist, AI & ML @UiPath
11:35 Power up your Robots: GenAI and GPT in REFramework (45')
Krzysztof Karaszewski, Global RPA Product Manager
12:20 🍕 Lunch Break (1hr)
13:20 From Concept to Quality: UiPath Test Suite for AI-powered Knowledge Bots (30')
Kamil Miśko, UiPath MVP, Senior RPA Developer @Zurich Insurance
13:50 Communications Mining - focus on AI capabilities (30')
Thomasz Wierzbicki, Business Analyst @Office Samurai
14:20 Polish MVP panel: Insights on MVP award achievements and career profiling
RPA In Healthcare Benefits, Use Case, Trend And Challenges 2024.pptxSynapseIndia
Your comprehensive guide to RPA in healthcare for 2024. Explore the benefits, use cases, and emerging trends of robotic process automation. Understand the challenges and prepare for the future of healthcare automation
Sustainability requires ingenuity and stewardship. Did you know Pigging Solutions pigging systems help you achieve your sustainable manufacturing goals AND provide rapid return on investment.
How? Our systems recover over 99% of product in transfer piping. Recovering trapped product from transfer lines that would otherwise become flush-waste, means you can increase batch yields and eliminate flush waste. From raw materials to finished product, if you can pump it, we can pig it.
INDIAN AIR FORCE FIGHTER PLANES LIST.pdfjackson110191
These fighter aircraft have uses outside of traditional combat situations. They are essential in defending India's territorial integrity, averting dangers, and delivering aid to those in need during natural calamities. Additionally, the IAF improves its interoperability and fortifies international military alliances by working together and conducting joint exercises with other air forces.
Paradigm Shifts in User Modeling: A Journey from Historical Foundations to Em...Erasmo Purificato
Slide of the tutorial entitled "Paradigm Shifts in User Modeling: A Journey from Historical Foundations to Emerging Trends" held at UMAP'24: 32nd ACM Conference on User Modeling, Adaptation and Personalization (July 1, 2024 | Cagliari, Italy)
5. H2O.ai Confidential
Retrieval Augmented Generation
● Answers are generated with additional
context from external knowledge
database.
● Grounding LLM answers to the
retrieved knowledge/facts.
● LLMs do not need to be retrained when
new knowledge is added or modified.
● H2OGPT / LangChain
Some Good Use-Cases
If you haven’t already started exploring
Summarization
● Business related summaries can be a
disruptive use-case. Every meeting of a
company can be recorded, transcribed,
and then summarized using LLM.
● Needs long-context models or iterative
reduction techniques.
● Intrinsically grounded LLM output.
Residual risk ok misinterpretations.
Creative Writing
● Promising area of interest for short
texts.
● Limited in creativity and novelty. Risk of
hallucination as it is challenging to
ground the generated text.
● LLM as a coworker/copilot.
6. v
H2O.ai Confidential
check out genai.h2o.ai for more inspiration
Function calling
Classification
Regression
Coding
Writing Helper
From simple spell and
grammar checking up to
professional rephrasing.
Broad applications when
LLMs are trained to call
APIs. Usage of external
tools (simple calculator
or browsing internet).
Translate natural
language to SQL queries
Use LLMs for classical
NLP use-cases such as
classification and
regression tasks. Can also
be retrieval augmented.
Writing code will be a
valuable task for LLMs
especially for small fine
grained tasks. This will
enable a lot more
software engineers to be
10x or the already 10x to
be 100x.
Autonomous
Agents
Fully automate repetitive
workflows. Make use of
function calling and
natural language
interaction. Requires LLM
memory management.
… Much More
There is more
Complexity
Function calling
Broad applications when
LLMs are trained to call
APIs. Usage of external
tools (simple calculator
or browsing internet).
Translate natural
language to SQL queries
Classification
Regression
Use LLMs for classical
NLP use-cases such as
classification and
regression tasks. Can also
be retrieval augmented.
7. H2O.ai Confidential
“How do I …?”
“Summarize!”
“What are …?”
Technical Ways to use
LLMs
Best practice Consider
Prompt engineering
Be very detailed in your instructions.
Use prompts with all relevant
information.
Ask for quality answer.
Refine your prompt.
Bias from prompt and from model.
Limited to train knowledge.
Hallucinations
Raw prompt engineering may
struggle in certain use-cases
Add context
Retrieve relevant context
Context placement.
Explore few shot prompting
Usually has a sweet spot for amount
of context (depends on use case,
start with 5 paragraphs / 512
characters each)
Fine-tuning
Build a proper evaluation first.
Start with small models.
Data is important.
Use H2O LLM Studio to quickly iterate
Needs training data.
Beware of fine-tuning for factuality.
Can be great for specific use-cases.
No prompt engineering needed.
RLHF
Model must be fine-tuned with in-
distribution data first.
PPO: requires a lot of high quality data
to train a good reward model.
DPO: loss is calculated directly on
accept/reject pairs.
Next token loss is replaced by a
scalar.
Very research oriented. Proven to be
useful for chat alignment.
PPO (with reward model) or DPO
(without reward model).
Increasing
Complexity
Always max out
the options above
(or already have
training data)
8. v
H2O.ai Confidential
Meetings
Expand Knowledge
Opportunities
Logs & Documents
● Identify work
duplication
● Find new synergies
● Improve staff
motivation by
providing up-to-date
information
● Personalized updates
on any related tasks
● Discover trends with
summaries on
summaries
LLM summarization
Business Intelligence
Automate Information Flow in Your Company
Demos & Presentations
11. H2O.ai Confidential
Solve with Prompt Engineering
In this talk, Andrew Ng discusses how AI can empower businesses of all sizes. He compares the rise of AI to the rise of literacy, explaining that just as literacy was once
limited to a few, AI is currently concentrated in the hands of highly skilled engineers at big tech companies. Ng believes that society can benefit from enabling everyone
to contribute to AI development. He explains that AI projects have been expensive to build, requiring skilled engineers and significant investment. Large tech companies
have been successful in making these investments pay off because they have millions or billions of users to generate revenue from their AI systems. However, this model
does not work for smaller businesses that do not have the same scale or resources. Ng gives examples of how AI can be beneficial to small businesses, such as a pizza
store owner using AI to improve sales by analyzing data and spotting patterns. He emphasizes that AI can work with modest amounts of data and that the real challenge
is the cost of hiring an AI team for small businesses. Ng proposes that the solution lies in new AI development platforms that focus on providing data rather than writing
code. These platforms allow individuals to upload data and train AI systems without extensive coding knowledge. He gives an example of an inspector using a platform to
teach an AI system how to detect defects in fabric. Ng acknowledges that these platforms are still evolving but believes they will become more accessible in the future.
He concludes by stating that democratizing access to AI is crucial for spreading wealth across society. Ng envisions a future where every accountant, store manager,
buyer, and quality inspector can build their own AI systems. He compares the impact of democratizing AI to the impact of widespread literacy, emphasizing the exciting
possibilities it holds for the future.
12. H2O.ai Confidential
Solve with Prompt Engineering
Sure! Here is a summary of Andrew Ng's talk "How AI could empower any business" in a single paragraph:
Andrew Ng discusses the current state of AI, where it is primarily controlled by a select few "high priests
and priestesses" in big tech companies, and how this limits the potential benefits of AI for society as a
whole. He argues that AI has the potential to create tremendous wealth, but only if it is democratized and
made accessible to everyone. To achieve this, he suggests that we need to shift the focus from asking
people to write lots of code to providing data, and highlights the emergence of new AI development
platforms that make this possible. He believes that empowering everyone to build AI systems will lead to
an incredibly exciting future, where small businesses and individuals can use AI to improve their operations
and create new opportunities.
13. H2O.ai Confidential
Solve with Prompt Engineering
Sure! Here's the summary of Andrew Ng's talk "How AI could empower any business" in a single paragraph:
Andrew Ng discusses the potential of AI to empower any business, not just large tech companies. He
highlights the long-tail of AI projects that are valuable but not being addressed, and the need for a
democratic approach to AI development. Ng introduces a new way to build AI systems that focuses on
providing data rather than writing code, making it easier for small businesses and individuals to participate.
He illustrates the potential of these platforms with examples of T-shirt demand forecasting and quality
control, and emphasizes the importance of empowering everyone to build AI systems for themselves to
create a more equitable distribution of wealth.
15. v
H2O.ai Confidential
Maximize baseline score
with prompt engineering
first. Choose maximum
tokens, that most or all
samples fit.
Prompt
Evaluate on sub-epochs
to see if your model
saturates quickly and
overfits. Manually inspect
the output and test some
hand-crafted samples.
Evaluate Often
Even with (Q)LoRA,
learning rate still has a
significant impact on the
training. Start with 0.0001
and AdamW.
Learning Rate
Spend quality time on
what the model is seeing.
Add context? Name, Title,
Text. Experiment with
order and delimiters.
Modify Prompt
Start with LoRA rank 4,
alpha 16, no-dropout.
Apply to all linear layers.
Decrease/increase LoRA
rank & alpha (2-4 * rank),
add a bit of dropout.
LoRA
Only very late, try to scale
up to larger models or
more (diverse) training
data. Also try to scale
down!
Scale Up
What do I Tune First?
So many hyperparameters
17. H2O.ai Confidential
In standard SFT training, Models are
rewarded for predicting the next
token. Factuality makes only a
miniscule fraction of the train loss.
Hallucinations Lack of Reasoning
02
01
They can easily fool us with eloquent
output.
They are only trained on text, and
don’t have deep understanding of
nature, math.
They don’t have any deeper goals or
desires.
…
Models inherit the bias from the
training corpus.
Models may struggle in classes
of applications, e.g. order
inversion, advanced math.
Models only have the
knowledge that they were
trained on (knowledge cutoff).
Models are susceptible to
prompt injection, “jailbreak”
attacks, data poisoning
attacks,…
Beyond Current Limits
A glimpse
18. v
H2O.ai Confidential
Hallucinations
LLMs are not a database search engine. Hallucinations (factual
incorrectness or made up context that “sounds good”) can be
very hard to spot.
By the architecture of LLMs, hallucinations can never be 100%
avoided but there are a few options to mitigate them:
RAG
Knowledge Graphs
Ground the answers by prompting with
high level abstractions and concepts first.
Provide Context
If very low rate of hallucinations is
important for your task, consider using the
best and largest models out there.
Larger/Better Models
Lower temperature or even greedy
decoding will produce more condensed
answers but they will be likely more
factual correct.
Low Temperature
Recommendations:
● Use in low-stakes applications,
combine with human oversight
● Source of inspiration suggestions
● Copilots over autonomous agents
Andrej Karpathy, State of GPT
19. v
H2O.ai Confidential
Reasoning
Fast, automatic, frequent, emotional,
stereotypic, unconscious.
- determine that an object is at a
greater distance than another
- localize the source of a specific
sound
- complete the phrase "war and ..."
- display disgust when seeing a
gruesome image
- solve 2+2=?
- read text on a billboard
- drive a car on an empty road
- understand simple sentences
System 1 (98 % of time)
Slow, effortful, infrequent, logical,
calculating, conscious.
- direct your attention towards the
clowns at the circus
- sustain a faster-than-normal
walking rate
- count the number of A's in a
certain text
- park into a tight parking space
- determine the price/quality ratio
of two washing machines
- determine the validity of a
complex logical reasoning
- solve 17 × 24
System 2 (2 % of time)
20. H2O.ai Confidential
LLMs are stochastic parrots
LLMs do not think, nor have consciousness.
But models can learn internal representations of
states as has been shown with Othello and Chess
(GPT-4 can play at around ELO 1800)
Do LLMs Have an Internal State?
21. v
H2O.ai Confidential
If the task is not solved intuitively …
… consider giving the model time (tokens) to
think about the answer.
Trick Models Into System 2
Chain-of-thought & Reflection
Instruct the model to go through it’s thoughts
step by step. Let the model reflect about its
own answer. While this has been proposed
quite a few times, I personally haven’t seen it
working well and the model is very biased
towards rewoking the original answer.
Decomposition
Decomposition of one task into multiple
smaller ones.
Step-back prompting
Retrieve background information and abstract
the question first. This grounds the answer to
fundamental concepts or principles.