Generative AI - Google Cloud Intro to LLMs

Beschreibung

Generative AI Karteikarten am Generative AI - Google Cloud Intro to LLMs, erstellt von Jake Kaldenbaugh am 14/07/2023.
Jake Kaldenbaugh
Karteikarten von Jake Kaldenbaugh, aktualisiert more than 1 year ago
Jake Kaldenbaugh
Erstellt von Jake Kaldenbaugh vor etwa ein Jahr
6
0

Zusammenfassung der Ressource

Frage Antworten
PaLM-E generalist robotics model from Google which transfers knowledge from visual+language models to robotics. supplements PaLM with ViT-22B which enables raw sensor data
How do LLMs work? LLMs represent text mathematically in a way that NNs can process by splitting text into tokes that encode subwords associated with high-dimensional vectors. Matrix mult is then applied to predict the next likely word.
LaMDA Language Model for Dialogue Applications: Built on Google's Transformer NN arch (2017), trained on dialogue. Incorporates "sensibleness", what makes sense in coversations.
Few-Shot vs Zero-Shot Training Few-shot: using sparse datasets to provide domain and task-specific performance. Implies that fine-tuning may not be required as few shot achieves similar results.
Bias in GPT-3 Occupations with higher edu req or phys labor were labled as Male 83% of 388. Race: Asian - high sentiment, Black - low. Religion:
What is Google Pathways? An attempt at a single model that can generalize across domains and tasks while being efficient. Pathways SystemL orchestrates dist comp for accelerators. Achieves good few-shot perf across lang+generation tasks
PaLM API & MakerSuite API: access to multi-turn, (content, chat) + gnrl purpose (summ, clssfic8n) MakerSuite: simplifies AI dev workflows - Tune model, synth data aug, genr8 embeddings, Resp+Sfty, Deploy & Scale
PEPT Parameter Efficient Prompt-Tuning: ft with a small dataset done in the prompt
Chain of Thought Prompting Prompt the model to break down its reasoning into steps "show work", creating more structured, organized and accurate responses. Most benefiicial for complex math & science.
Computer Vision Trends in 2020 began moving from Convolution NNs to Transformers arch.
GANs Generative Adversarial Networks: set up two opposing models: generator+discriminator, use results to improve ability to "win"
How do Diffusion Models Work? Systematically/slowly destroy structure in a data distribution through iterative fwd diffusion, the learn reverse diffusion to restore data.
CALM Confident Adaptive Language Modeling: accelerating LM text generation by improving efficiency at inference time because some word predictions are higher prob than others. CLAM dynamically distributes the computational effort across gen timesteps
In-Context Learning (Few-Shot Prompting) Ability for LLMs to do tasks after seeing a few examples. LLMs contain smaller simpler linear models based on examples within massive training sets. NNs perhaps have internal ML models.
Zusammenfassung anzeigen Zusammenfassung ausblenden

ähnlicher Inhalt

Klausurfragen aktuell
Anne Malten
Tag der deutschen Einheit
barbara91
KPOL-Fragen
Lisa-Maria Hauschild
Kabale und Liebe - Umfangreiche Lernfolien
Laura Overhoff
Gesetzliche Sozialversicherung
Maximilian Mustermann
1_Blut und Blutkreislauf
Ulrike Renauer
13. Altklausuren - Allgemeinwissen Makro
Rafael Mentges
Geometrie 33-48
Christoph Affolter
AVO 2017
Schmolli Schmoll
Vetie: spez Patho 2018
Johanna Tr
Vetie Berufsrecht 2020 !
Alena Gesing