Skip to content

Let AI think and express like you. This framework provides a complete assembly line: from the noise processing of original chat records, to the seamless switching of multi-model adaptation layers, to local lightweight fine-tuning (LoRA).

Notifications You must be signed in to change notification settings

whyzhow/PersonalStyleAI-Framework-

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

10 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

PersonalStyleAI-Framework-

Let AI think and express like you. This framework provides a complete assembly line: from the noise processing of original chat records, to the seamless switching of multi-model adaptation layers, to local lightweight fine-tuning (LoRA).

🧭 Working principle (Workflow)

This project divides the personalisation process of AI into three core stages:

  1. Data Alchemy (Extract):

Convert messy chat records (including emojis, links, spam) into high-quality conversation pairs.

  1. Adaptor Centre (Adapt):

Through a unified interface, your personal style can be easily loaded to GPT-4, Claude or local Llama 3.

  1. Style Evolution (Evolve):

Use fine-tuning technology to solidify your language habits in the weight of the model, not just rely on Prompt.


πŸ“Š Example of data conversion

Input (original chat.txt):

User A: Have you eaten yet?

I: Eat, [expression] I'm really hungry, https://link.com

User A: Then let's go.

Me: Indeed, Let'S Go.

Output (JSONL after cleaning):

{

"Messages": [

{"Role": "user", "content": "Have you eaten yet?"},

{"Role": "assistant", "content": "I'm really hungry after eating"}

]

}

πŸ“‚ Project structure


PersonalStyleAI-Framework/

β”œβ”€β”€ data/

β”‚ β”œβ”€β”€ raw/ # Original chat records (such as chat.txt)

β”‚ └── processed/ # JSONL training data set after cleaning

β”œβ”€β”€ src/ # Source code

β”‚ β”œβ”€β”€ core/ # adaptor logic and factory mode implementation

β”‚ β”œβ”€β”€ utils/ # Data preprocessing and string cleaning tools

β”‚ └── trainers/ # Model fine-tuning script (based on PEFT/LoRA)

β”œβ”€β”€ pyproject.toml # Modern Python Dependency and Project Configuration

β”œβ”€β”€ preprocess_data.py # Data processing entry script

β”œβ”€β”€ main.py # Style Dialogue Test Entrance

└── .env.example # Environment variable template

πŸ“‚ Detailed description of the project module

  1. Core adaptor (src/core/)

Adopt factory model design. This means that if you want to switch from OpenAI to local Ollama, you only need to change the one-line configuration without rewriting the business logic.

  1. Cleaning toolbox (src/utils/)

Efficient regular expressions are preset and optimised for text exported by social software.

  1. Environmental isolation

Use .env to manage sensitive information and manage dependency hierarchy through pyproject.toml.

πŸš€ Quick installation

1.Basic version (only call API)

# Clone Project

Git clone [https://github.com/your username/PersonalStyleAI-Framework.git](https://github.com/your username/Personal StyleAI-Framework.git)

Cd PersonalStyleAI-Framework

# Create a virtual environment and install core dependencies

Python -m venv venv

Source venv/bin/activate # Windows use venv\Scripts\activate

Pip install-e.
  1. Configure the key

Create a .env file and fill in your API Key:

cp .env.example .env
  1. Build your style

Collect data: Put your chat records or articles into data/raw/chat.txt. Running cleaning:

python preprocess_data.py

The script will generate data/processed/train.jsonl, which is a "textbook" for AI to learn your style. 4. Run the dialogue

python main.py

πŸ›  Advanced: local fine-tuning

If you have a graphics card that supports CUDA, you can install fine-tuning components for local training: Pip install -e ".[train]"

# Run the fine-tuning script (need to configure parameters according to src/trainers)

Python run_train.py

🀝 Contribution and feedback

If you have any suggestions for improvement or want to add more AI adaptors (such as Anthropic or DeepSeek), welcome to submit a Pull Request or open an Issue discussion.

About

Let AI think and express like you. This framework provides a complete assembly line: from the noise processing of original chat records, to the seamless switching of multi-model adaptation layers, to local lightweight fine-tuning (LoRA).

Topics

Resources

Stars

Watchers

Forks

Languages