|
--- |
|
language: en |
|
datasets: |
|
- fka/awesome-chatgpt-prompts |
|
tags: |
|
- llama-2 |
|
- fine-tuning |
|
- causal-lm |
|
- chatgpt |
|
license: apache-2.0 |
|
model_name: fine-tuned-llama2 |
|
widget: |
|
- text: "I want you to act as a Linux terminal." |
|
- text: "Imagine you are an Ethereum Developer tasked with creating a smart contract." |
|
--- |
|
|
|
# Fine-Tuned Llama 2 Model |
|
|
|
## Model Description |
|
|
|
This model is a fine-tuned version of Llama 2, trained on a dataset of diverse prompts and scenarios. The model has been designed to generate responses based on various tasks described in the `prompt` column of the dataset. The fine-tuning process aims to improve the model's performance in handling specific tasks across multiple domains, such as software development, SEO, and Linux commands. |
|
|
|
## Dataset Information |
|
|
|
The dataset used for fine-tuning this model consists of two primary columns: |
|
|
|
1. **`act`**: The role or scenario that the model is asked to act upon. For example: |
|
- "An Ethereum Developer" |
|
- "SEO Prompt" |
|
- "Linux Terminal" |
|
|
|
2. **`prompt`**: The detailed task or scenario description related to the `act`. This provides context and specific instructions that the model needs to follow. Example prompts: |
|
- *"Imagine you are an experienced Ethereum developer tasked with creating a smart contract for a blockchain messenger..."* |
|
- *"Using WebPilot, create an outline for an article that will be 2,000 words on the keyword 'Best SEO prompts'..."* |
|
- *"I want you to act as a linux terminal. I will type commands and you will reply with what the terminal should show..."* |
|
|
|
The dataset includes a wide range of scenarios aimed at helping the model generalize across technical and creative tasks. |
|
|
|
### Dataset Samples |
|
|
|
| **Act** | **Prompt** | |
|
|--------------------|-------------------------------------------------------------------------------------------------------------| |
|
| Ethereum Developer | Imagine you are an experienced Ethereum developer tasked with creating a smart contract for a blockchain... | |
|
| SEO Prompt | Using WebPilot, create an outline for an article that will be 2,000 words on the keyword 'Best SEO prompts'...| |
|
| Linux Terminal | I want you to act as a linux terminal. I will type commands and you will reply with what the terminal should show...| |
|
|
|
## Output Example |
|
|
|
The model has been fine-tuned to generate detailed, contextually relevant responses based on the prompts provided. Here’s an example of how the model might respond to a sample prompt: |
|
|
|
### Input: |
|
**Act**: *Linux Terminal* |
|
|
|
### Output: |
|
**Prompt**: |
|
*I want you to act as a Linux terminal. I will type commands, and you will reply with what the terminal should show. Execute the command `ls`.* |
|
|
|
In this scenario, the model understands that it should act as a Linux terminal and simulate the result of running the `ls` command. |
|
|
|
### Another Example: |
|
|
|
### Input: |
|
**Act**: *Ethereum Developer* |
|
|
|
### Output: |
|
**Prompt**: |
|
*Imagine you are an experienced Ethereum developer tasked with creating a smart contract for a blockchain messenger...* |
|
|
|
In this example, the model generates Solidity code based on the prompt, addressing the requirements for a blockchain messenger. |
|
|
|
## How to Use the Model |
|
|
|
This model can be loaded and used through the Hugging Face Hub: |
|
|
|
```python |
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
|
|
# Load the model and tokenizer |
|
tokenizer = AutoTokenizer.from_pretrained("Manish-KT/Fine_tune_Llama_2") |
|
model = AutoModelForCausalLM.from_pretrained("Manish-KT/Fine_tune_Llama_2") |
|
|
|
# Encode the prompt |
|
inputs = tokenizer("I want you to act as a linux terminal.", return_tensors="pt") |
|
|
|
# Generate the response |
|
outputs = model.generate(inputs["input_ids"], max_length=100) |
|
print(tokenizer.decode(outputs[0], skip_special_tokens=True)) |
|
``` |
|
|
|
## Acknowledgements |
|
|
|
Special thanks to the creators of the dataset `fka/awesome-chatgpt-prompts`, which provided the rich prompts and diverse scenarios used in fine-tuning this model. |
|
|
|
## License |
|
This model is open-sourced and can be used for both commercial and non-commercial purposes. Please ensure that you attribute the original dataset and respect any usage policies. |
|
--- |
|
license: mit |
|
--- |
|
|