Glossary
Keyword | Reasoning |
---|---|
Model Fintuning | Model fine-tuning adapts a pre-trained machine learning model for a related task. Instead of training from scratch, this method uses a model that has already learned useful features from another task and fine-tunes it on a new dataset or task. |
Model Inference | The model has already been trained and evaluated, and has been deployed into the actual target hardware. Data is input into the model, and the model predicts the results. |
Large Language Model/LLM | A type of artificial intelligence model aimed at understanding and generating human language. Trained on a large amount of text data, it can perform a wide range of tasks, including text summarization, translation, sentiment analysis, and so on. |
Federated learning | A new paradigm in machine learning aims to train centralized models while keeping a large amount of client-side training data from being uploaded to avoid data leaks. |
Deep Gradient Compression/DGC | Reduce the network bandwidth required for exchanging gradient information |
Gradient inversion attack | A specialized form of adversarial attack aimed at systems utilizing Federated Learning. In this type of attack, adversaries attempt to 'invert' the aggregated gradient updates sent to the centralized server. By doing so, they seek to reconstruct or infer sensitive information from the original data that contributed to these gradient updates. This poses a significant risk to data exploit and security within Federated Learning ecosystems. |
Byzantine attack |
|
RAG | Retrieval-Augmented Generation (RAG) is a method that improves the responses of large language models (like GPT-3) by letting them check an external, trustworthy knowledge base before they answer. This means they can give more accurate and up-to-date answers, especially in specialized areas or using a company's own data, without having to be retrained. This makes these models even better at jobs like answering questions, translating languages, and completing sentences, and it's a budget-friendly way to make sure they stay useful and correct in many different situations. |
Last updated