| 12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849 |
- In the context of LLMs, what is "Attention"?
- In the context of LLMs, what is a completion?
- In the context of LLMs, what is a prompt?
- In the context of LLMs, what is GELU?
- In the context of LLMs, what is RELU?
- In the context of LLMs, what is softmax?
- In the context of LLMs, what is decoding?
- In the context of LLMs, what is encoding?
- In the context of LLMs, what is tokenizing?
- In the context of LLMs, what is an embedding?
- In the context of LLMs, what is quantization?
- In the context of LLMs, what is a tensor?
- In the context of LLMs, what is a sparse tensor?
- In the context of LLMs, what is a vector?
- In the context of LLMs, how is attention implemented?
- In the context of LLMs, why is attention all you need?
- In the context of LLMs, what is "RoPe" and what is it used for?
- In the context of LLMs, what is "LoRA" and what is it used for?
- In the context of LLMs, what are weights?
- In the context of LLMs, what are biases?
- In the context of LLMs, what are checkpoints?
- In the context of LLMs, what is "perplexity"?
- In the context of LLMs, what are models?
- In the context of machine-learning, what is "catastrophic forgetting"?
- In the context of machine-learning, what is "elastic weight consolidation (EWC)"?
- In the context of neural nets, what is a hidden layer?
- In the context of neural nets, what is a convolution?
- In the context of neural nets, what is dropout?
- In the context of neural nets, what is cross-entropy?
- In the context of neural nets, what is over-fitting?
- In the context of neural nets, what is under-fitting?
- What is the difference between an interpreted computer language and a compiled computer language?
- In the context of software development, what is a debugger?
- When processing using a GPU, what is off-loading?
- When processing using a GPU, what is a batch?
- When processing using a GPU, what is a block?
- When processing using a GPU, what is the difference between a batch and a block?
- When processing using a GPU, what is a scratch tensor?
- When processing using a GPU, what is a layer?
- When processing using a GPU, what is a cache?
- When processing using a GPU, what is unified memory?
- When processing using a GPU, what is VRAM?
- When processing using a GPU, what is a kernel?
- When processing using a GPU, what is "metal"?
- In the context of LLMs, what are "Zero-Shot", "One-Shot" and "Few-Shot" learning models?
- In the context of LLMs, what is the "Transformer-model" architecture?
- In the context of LLMs, what is "Multi-Head Attention"?
- In the context of LLMs, what is "Self-Attention"?
- In the context of transformer-model architectures, how do attention mechanisms use masks?
|