repo
stringlengths
7
90
file_url
stringlengths
81
315
file_path
stringlengths
4
228
content
stringlengths
0
32.8k
language
stringclasses
1 value
license
stringclasses
7 values
commit_sha
stringlengths
40
40
retrieved_at
stringdate
2026-01-04 14:38:15
2026-01-05 02:33:18
truncated
bool
2 classes
browser-use/browser-use
https://github.com/browser-use/browser-use/blob/630f85dd05127c9d42810a5db235a14f5bac9043/examples/custom-functions/file_upload.py
examples/custom-functions/file_upload.py
""" Example of implementing file upload functionality. This shows how to upload files to file input elements on web pages. """ import asyncio import logging import os import sys import anyio sys.path.append(os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))) from dotenv import load_dotenv load_dotenv() from browser_use import ChatOpenAI from browser_use.agent.service import Agent, Tools from browser_use.agent.views import ActionResult from browser_use.browser import BrowserSession from browser_use.browser.events import UploadFileEvent logger = logging.getLogger(__name__) # Initialize tools tools = Tools() @tools.action('Upload file to interactive element with file path') async def upload_file(index: int, path: str, browser_session: BrowserSession, available_file_paths: list[str]): if path not in available_file_paths: return ActionResult(error=f'File path {path} is not available') if not os.path.exists(path): return ActionResult(error=f'File {path} does not exist') try: # Get the DOM element by index dom_element = await browser_session.get_dom_element_by_index(index) if dom_element is None: msg = f'No element found at index {index}' logger.info(msg) return ActionResult(error=msg) # Check if it's a file input element if dom_element.tag_name.lower() != 'input' or dom_element.attributes.get('type') != 'file': msg = f'Element at index {index} is not a file input element' logger.info(msg) return ActionResult(error=msg) # Dispatch the upload file event event = browser_session.event_bus.dispatch(UploadFileEvent(node=dom_element, file_path=path)) await event msg = f'Successfully uploaded file to index {index}' logger.info(msg) return ActionResult(extracted_content=msg, include_in_memory=True) except Exception as e: msg = f'Failed to upload file to index {index}: {str(e)}' logger.info(msg) return ActionResult(error=msg) async def main(): """Main function to run the example""" browser_session = BrowserSession() await browser_session.start() llm = ChatOpenAI(model='gpt-4.1-mini') # List of file paths the agent is allowed to upload # In a real scenario, you'd want to be very careful about what files # the agent can access and upload available_file_paths = [ '/tmp/test_document.pdf', '/tmp/test_image.jpg', ] # Create test files if they don't exist for file_path in available_file_paths: if not os.path.exists(file_path): await anyio.Path(file_path).write_text('Test file content for upload example') # Create the agent with file upload capability agent = Agent( task=""" Go to https://www.w3schools.com/howto/howto_html_file_upload_button.asp and try to upload one of the available test files. """, llm=llm, browser_session=browser_session, tools=tools, # Pass the available file paths to the tools context custom_context={'available_file_paths': available_file_paths}, ) # Run the agent await agent.run(max_steps=10) # Cleanup await browser_session.kill() # Clean up test files for file_path in available_file_paths: if os.path.exists(file_path): os.remove(file_path) if __name__ == '__main__': asyncio.run(main())
python
MIT
630f85dd05127c9d42810a5db235a14f5bac9043
2026-01-04T14:38:16.467592Z
false
browser-use/browser-use
https://github.com/browser-use/browser-use/blob/630f85dd05127c9d42810a5db235a14f5bac9043/examples/custom-functions/action_filters.py
examples/custom-functions/action_filters.py
""" Action filters (domains) let you limit actions available to the Agent on a step-by-step/page-by-page basis. @registry.action(..., domains=['*']) async def some_action(browser_session: BrowserSession): ... This helps prevent the LLM from deciding to use an action that is not compatible with the current page. It helps limit decision fatigue by scoping actions only to pages where they make sense. It also helps prevent mis-triggering stateful actions or actions that could break other programs or leak secrets. For example: - only run on certain domains @registry.action(..., domains=['example.com', '*.example.com', 'example.co.*']) (supports globs, but no regex) - only fill in a password on a specific login page url - only run if this action has not run before on this page (e.g. by looking up the url in a file on disk) During each step, the agent recalculates the actions available specifically for that page, and informs the LLM. """ import asyncio import os import sys sys.path.append(os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))) from dotenv import load_dotenv load_dotenv() from browser_use import ChatOpenAI from browser_use.agent.service import Agent, Tools from browser_use.browser import BrowserSession # Initialize tools and registry tools = Tools() registry = tools.registry # Action will only be available to Agent on Google domains because of the domain filter @registry.action(description='Trigger disco mode', domains=['google.com', '*.google.com']) async def disco_mode(browser_session: BrowserSession): # Execute JavaScript using CDP cdp_session = await browser_session.get_or_create_cdp_session() await cdp_session.cdp_client.send.Runtime.evaluate( params={ 'expression': """(() => { // define the wiggle animation document.styleSheets[0].insertRule('@keyframes wiggle { 0% { transform: rotate(0deg); } 50% { transform: rotate(10deg); } 100% { transform: rotate(0deg); } }'); document.querySelectorAll("*").forEach(element => { element.style.animation = "wiggle 0.5s infinite"; }); })()""" }, session_id=cdp_session.session_id, ) # Custom filter function that checks URL async def is_login_page(browser_session: BrowserSession) -> bool: """Check if current page is a login page.""" try: # Get current URL using CDP cdp_session = await browser_session.get_or_create_cdp_session() result = await cdp_session.cdp_client.send.Runtime.evaluate( params={'expression': 'window.location.href', 'returnByValue': True}, session_id=cdp_session.session_id ) url = result.get('result', {}).get('value', '') return 'login' in url.lower() or 'signin' in url.lower() except Exception: return False # Note: page_filter is not directly supported anymore, so we'll just use domains # and check the condition inside the function @registry.action(description='Use the force, luke', domains=['*']) async def use_the_force(browser_session: BrowserSession): # Check if it's a login page if not await is_login_page(browser_session): return # Skip if not a login page # Execute JavaScript using CDP cdp_session = await browser_session.get_or_create_cdp_session() await cdp_session.cdp_client.send.Runtime.evaluate( params={ 'expression': """(() => { document.querySelector('body').innerHTML = 'These are not the droids you are looking for'; })()""" }, session_id=cdp_session.session_id, ) async def main(): """Main function to run the example""" browser_session = BrowserSession() await browser_session.start() llm = ChatOpenAI(model='gpt-4.1-mini') # Create the agent agent = Agent( # disco mode will not be triggered on apple.com because the LLM won't be able to see that action available, it should work on Google.com though. task=""" Go to apple.com and trigger disco mode (if dont know how to do that, then just move on). Then go to google.com and trigger disco mode. After that, go to the Google login page and Use the force, luke. """, llm=llm, browser_session=browser_session, tools=tools, ) # Run the agent await agent.run(max_steps=10) # Cleanup await browser_session.kill() if __name__ == '__main__': asyncio.run(main())
python
MIT
630f85dd05127c9d42810a5db235a14f5bac9043
2026-01-04T14:38:16.467592Z
false
browser-use/browser-use
https://github.com/browser-use/browser-use/blob/630f85dd05127c9d42810a5db235a14f5bac9043/examples/custom-functions/parallel_agents.py
examples/custom-functions/parallel_agents.py
""" Simple parallel multi-agent example. This launches multiple agents in parallel to work on different tasks simultaneously. No complex orchestrator - just direct parallel execution. @file purpose: Demonstrates parallel multi-agent execution using asyncio """ import asyncio import os import sys sys.path.append(os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))) from dotenv import load_dotenv load_dotenv() from browser_use import Agent from browser_use.llm.google import ChatGoogle # ============================================================================ # πŸ”§ SIMPLE CONFIGURATION - CHANGE THIS TO YOUR DESIRED TASK # ============================================================================ MAIN_TASK = 'find age of ronaldo and messi' # Simple test - let's start with just one person to see what happens # MAIN_TASK = "find age of elon musk" # ============================================================================ async def create_subtasks(main_task: str, llm) -> list[str]: """ Use LLM to break down main task into logical subtasks Real examples of how this works: Input: "what is the revenue of nvidia, microsoft, tesla" Output: [ "Find Nvidia's current revenue and financial data", "Find Microsoft's current revenue and financial data", "Find Tesla's current revenue and financial data" ] Input: "what are ages of musk, altman, bezos, gates" Output: [ "Find Elon Musk's age and birth date", "Find Sam Altman's age and birth date", "Find Jeff Bezos's age and birth date", "Find Bill Gates's age and birth date" ] Input: "what is the population of tokyo, new york, london, paris" Output: [ "Find Tokyo's current population", "Find New York's current population", "Find London's current population", "Find Paris's current population" ] Input: "name top 10 yc companies by revenue" Output: [ "Research Y Combinator's top companies by revenue", "Find revenue data for top YC companies", "Compile list of top 10 YC companies by revenue" ] """ prompt = f""" Break down this main task into individual, separate subtasks where each subtask focuses on ONLY ONE specific person, company, or item: Main task: {main_task} RULES: - Each subtask must focus on ONLY ONE person/company/item - Do NOT combine multiple people/companies/items in one subtask - Each subtask should be completely independent - If the main task mentions multiple items, create one subtask per item Return only the subtasks, one per line, without numbering or bullets. Each line should focus on exactly ONE person/company/item. """ try: # Use the correct method for ChatGoogle response = await llm.ainvoke(prompt) # Debug: Print the response type and content print(f'DEBUG: Response type: {type(response)}') print(f'DEBUG: Response content: {response}') # Handle different response types - ChatGoogle returns string content if hasattr(response, 'content'): content = response.content elif isinstance(response, str): content = response elif hasattr(response, 'text'): content = response.text else: # Convert to string if it's some other type content = str(response) # Split by newlines and clean up subtasks = [task.strip() for task in content.strip().split('\n') if task.strip()] # Remove any numbering or bullets that the LLM might add cleaned_subtasks = [] for task in subtasks: # Remove common prefixes like "1. ", "- ", "* ", etc. cleaned = task.lstrip('0123456789.-* ') if cleaned: cleaned_subtasks.append(cleaned) return cleaned_subtasks if cleaned_subtasks else simple_split_task(main_task) except Exception as e: print(f'Error creating subtasks: {e}') # Fallback to simple split return simple_split_task(main_task) def simple_split_task(main_task: str) -> list[str]: """Simple fallback: split task by common separators""" task_lower = main_task.lower() # Try to split by common separators if ' and ' in task_lower: parts = main_task.split(' and ') return [part.strip() for part in parts if part.strip()] elif ', ' in main_task: parts = main_task.split(', ') return [part.strip() for part in parts if part.strip()] elif ',' in main_task: parts = main_task.split(',') return [part.strip() for part in parts if part.strip()] # If no separators found, return the original task return [main_task] async def run_single_agent(task: str, llm, agent_id: int) -> tuple[int, str]: """Run a single agent and return its result""" print(f'πŸš€ Agent {agent_id} starting: {task}') print(f' πŸ“ This agent will focus ONLY on: {task}') print(f' 🌐 Creating isolated browser instance for agent {agent_id}') try: # Create agent with its own browser session (separate browser instance) import tempfile from browser_use.browser import BrowserSession from browser_use.browser.profile import BrowserProfile # Create a unique temp directory for this agent's browser data temp_dir = tempfile.mkdtemp(prefix=f'browser_agent_{agent_id}_') # Create browser profile with custom user data directory and single tab focus profile = BrowserProfile() profile.user_data_dir = temp_dir profile.headless = False # Set to True if you want headless mode profile.keep_alive = False # Don't keep browser alive after task # Add custom args to prevent new tabs and popups profile.args = [ '--disable-popup-blocking', '--disable-extensions', '--disable-plugins', '--disable-images', # Faster loading '--no-first-run', '--disable-default-apps', '--disable-background-timer-throttling', '--disable-backgrounding-occluded-windows', '--disable-renderer-backgrounding', ] # Create a new browser session for each agent with the custom profile browser_session = BrowserSession(browser_profile=profile) # Debug: Check initial tab count try: await browser_session.start() initial_tabs = await browser_session._cdp_get_all_pages() print(f' πŸ“Š Agent {agent_id} initial tab count: {len(initial_tabs)}') except Exception as e: print(f' ⚠️ Could not check initial tabs for agent {agent_id}: {e}') # Create agent with the dedicated browser session and disable auto URL detection agent = Agent(task=task, llm=llm, browser_session=browser_session, preload=False) # Run the agent with timeout to prevent hanging try: result = await asyncio.wait_for(agent.run(), timeout=300) # 5 minute timeout except TimeoutError: print(f'⏰ Agent {agent_id} timed out after 5 minutes') result = 'Task timed out' # Debug: Check final tab count try: final_tabs = await browser_session._cdp_get_all_pages() print(f' πŸ“Š Agent {agent_id} final tab count: {len(final_tabs)}') for i, tab in enumerate(final_tabs): print(f' Tab {i + 1}: {tab.get("url", "unknown")[:50]}...') except Exception as e: print(f' ⚠️ Could not check final tabs for agent {agent_id}: {e}') # Extract clean result from the agent history clean_result = extract_clean_result(result) # Close the browser session for this agent try: await browser_session.kill() except Exception as e: print(f'⚠️ Warning: Error closing browser for agent {agent_id}: {e}') print(f'βœ… Agent {agent_id} completed and browser closed: {task}') return agent_id, clean_result except Exception as e: error_msg = f'Agent {agent_id} failed: {str(e)}' print(f'❌ {error_msg}') return agent_id, error_msg def extract_clean_result(agent_result) -> str: """Extract clean result from agent history""" try: # Get the last result from the agent history if hasattr(agent_result, 'all_results') and agent_result.all_results: last_result = agent_result.all_results[-1] if hasattr(last_result, 'extracted_content') and last_result.extracted_content: return last_result.extracted_content # Fallback to string representation return str(agent_result) except Exception: return 'Result extraction failed' async def run_parallel_agents(): """Run multiple agents in parallel on different tasks""" # Use Gemini 1.5 Flash llm = ChatGoogle(model='gemini-1.5-flash') # Main task to break down - use the simple configuration main_task = MAIN_TASK print(f'🎯 Main task: {main_task}') print('🧠 Creating subtasks using LLM...') # Create subtasks using LLM subtasks = await create_subtasks(main_task, llm) print(f'πŸ“‹ Created {len(subtasks)} subtasks:') for i, task in enumerate(subtasks, 1): print(f' {i}. {task}') print(f'\nπŸ”₯ Starting {len(subtasks)} agents in parallel...') print('πŸ” Each agent will get its own browser instance with exactly ONE tab') print(f'πŸ“Š Expected: {len(subtasks)} browser instances, {len(subtasks)} tabs total') # Create tasks for parallel execution agent_tasks = [run_single_agent(task, llm, i + 1) for i, task in enumerate(subtasks)] # Run all agents in parallel using asyncio.gather results = await asyncio.gather(*agent_tasks) # Print results print('\n' + '=' * 60) print('πŸ“Š PARALLEL EXECUTION RESULTS') print('=' * 60) for agent_id, result in results: print(f'\nπŸ€– Agent {agent_id} result:') print(f'Task: {subtasks[agent_id - 1]}') print(f'Result: {result}') print('-' * 50) print(f'\nπŸŽ‰ All {len(subtasks)} parallel agents completed!') def main(): """Main function to run parallel agents""" # Check if Google API key is available api_key = os.getenv('GOOGLE_API_KEY') if not api_key: print('❌ Error: GOOGLE_API_KEY environment variable not set') print('Please set your Google API key to use parallel agents') print('You can set it with: export GOOGLE_API_KEY="your-key-here"') sys.exit(1) # Check if API key looks valid (Google API keys are typically 39 characters) if len(api_key) < 20: print(f'⚠️ Warning: GOOGLE_API_KEY seems too short ({len(api_key)} characters)') print('Google API keys are typically 39 characters long') print('Continuing anyway, but this might cause authentication issues...') print('πŸš€ Starting parallel multi-agent example...') print(f'πŸ“ Task: {MAIN_TASK}') print('This will dynamically create agents based on task complexity') print('-' * 60) asyncio.run(run_parallel_agents()) if __name__ == '__main__': main()
python
MIT
630f85dd05127c9d42810a5db235a14f5bac9043
2026-01-04T14:38:16.467592Z
false
browser-use/browser-use
https://github.com/browser-use/browser-use/blob/630f85dd05127c9d42810a5db235a14f5bac9043/examples/custom-functions/save_to_file_hugging_face.py
examples/custom-functions/save_to_file_hugging_face.py
import asyncio import os import sys sys.path.append(os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))) from dotenv import load_dotenv load_dotenv() from pydantic import BaseModel from browser_use import ChatOpenAI from browser_use.agent.service import Agent from browser_use.tools.service import Tools # Initialize tools first tools = Tools() class Model(BaseModel): title: str url: str likes: int license: str class Models(BaseModel): models: list[Model] @tools.action('Save models', param_model=Models) def save_models(params: Models): with open('models.txt', 'a') as f: for model in params.models: f.write(f'{model.title} ({model.url}): {model.likes} likes, {model.license}\n') # video: https://preview.screen.studio/share/EtOhIk0P async def main(): task = 'Look up models with a license of cc-by-sa-4.0 and sort by most likes on Hugging face, save top 5 to file.' model = ChatOpenAI(model='gpt-4.1-mini') agent = Agent(task=task, llm=model, tools=tools) await agent.run() if __name__ == '__main__': asyncio.run(main())
python
MIT
630f85dd05127c9d42810a5db235a14f5bac9043
2026-01-04T14:38:16.467592Z
false
browser-use/browser-use
https://github.com/browser-use/browser-use/blob/630f85dd05127c9d42810a5db235a14f5bac9043/examples/custom-functions/actor_use.py
examples/custom-functions/actor_use.py
import asyncio import os import sys from browser_use.browser.session import BrowserSession sys.path.append(os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))) from dotenv import load_dotenv load_dotenv() from browser_use import ActionResult, Agent, ChatOpenAI, Tools tools = Tools() llm = ChatOpenAI(model='gpt-4.1-mini') @tools.registry.action('Click on submit button') async def click_submit_button(browser_session: BrowserSession): page = await browser_session.must_get_current_page() submit_button = await page.must_get_element_by_prompt('submit button', llm) await submit_button.click() return ActionResult(is_done=True, extracted_content='Submit button clicked!') async def main(): task = 'go to brower-use.com and then click on the submit button' agent = Agent(task=task, llm=llm, tools=tools) await agent.run() if __name__ == '__main__': asyncio.run(main())
python
MIT
630f85dd05127c9d42810a5db235a14f5bac9043
2026-01-04T14:38:16.467592Z
false
browser-use/browser-use
https://github.com/browser-use/browser-use/blob/630f85dd05127c9d42810a5db235a14f5bac9043/examples/custom-functions/2fa.py
examples/custom-functions/2fa.py
import os import sys sys.path.append(os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))) from dotenv import load_dotenv load_dotenv() from browser_use import Agent secret_key = os.environ.get('OTP_SECRET_KEY') if not secret_key: # For this example copy the code from the website https://authenticationtest.com/totpChallenge/ # For real 2fa just copy the secret key when you setup 2fa, you can get this e.g. in 1Password secret_key = 'JBSWY3DPEHPK3PXP' sensitive_data = {'bu_2fa_code': secret_key} task = """ 1. Go to https://authenticationtest.com/totpChallenge/ and try to log in. 2. If prompted for 2FA code: Input the the secret bu_2fa_code. When you input bu_2fa_code, the 6 digit code will be generated automatically. """ Agent(task=task, sensitive_data=sensitive_data).run_sync() # type: ignore
python
MIT
630f85dd05127c9d42810a5db235a14f5bac9043
2026-01-04T14:38:16.467592Z
false
browser-use/browser-use
https://github.com/browser-use/browser-use/blob/630f85dd05127c9d42810a5db235a14f5bac9043/examples/custom-functions/advanced_search.py
examples/custom-functions/advanced_search.py
import asyncio import http.client import json import os import sys sys.path.append(os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))) from dotenv import load_dotenv load_dotenv() import logging from pydantic import BaseModel from browser_use import ActionResult, Agent, ChatOpenAI, Tools from browser_use.browser.profile import BrowserProfile logger = logging.getLogger(__name__) class Person(BaseModel): name: str email: str | None = None class PersonList(BaseModel): people: list[Person] SERP_API_KEY = os.getenv('SERPER_API_KEY') if not SERP_API_KEY: raise ValueError('SERPER_API_KEY is not set') tools = Tools(exclude_actions=['search'], output_model=PersonList) @tools.registry.action('Search the web for a specific query. Returns a short description and links of the results.') async def search_web(query: str): # do a serp search for the query conn = http.client.HTTPSConnection('google.serper.dev') payload = json.dumps({'q': query}) headers = {'X-API-KEY': SERP_API_KEY, 'Content-Type': 'application/json'} conn.request('POST', '/search', payload, headers) res = conn.getresponse() data = res.read() serp_data = json.loads(data.decode('utf-8')) # exclude searchParameters and credits serp_data = {k: v for k, v in serp_data.items() if k not in ['searchParameters', 'credits']} # keep the value of the key "organic" organic = serp_data.get('organic', []) # remove the key "position" organic = [{k: v for k, v in d.items() if k != 'position'} for d in organic] # print the original data logger.debug(json.dumps(organic, indent=2)) # to string organic_str = json.dumps(organic) return ActionResult(extracted_content=organic_str, include_in_memory=False, include_extracted_content_only_once=True) names = [ 'Ruedi Aebersold', 'Bernd Bodenmiller', 'Eugene Demler', 'Erich Fischer', 'Pietro Gambardella', 'Matthias Huss', 'Reto Knutti', 'Maksym Kovalenko', 'Antonio Lanzavecchia', 'Maria Lukatskaya', 'Jochen Markard', 'Javier PΓ©rez-RamΓ­rez', 'Federica Sallusto', 'Gisbert Schneider', 'Sonia I. Seneviratne', 'Michael Siegrist', 'Johan Six', 'Tanja Stadler', 'Shinichi Sunagawa', 'Michael Bruce Zimmermann', ] async def main(): task = 'use search_web with "find email address of the following ETH professor:" for each of the following persons in a list of actions. Finally return the list with name and email if provided - do always 5 at once' task += '\n' + '\n'.join(names) model = ChatOpenAI(model='gpt-4.1-mini') browser_profile = BrowserProfile() agent = Agent(task=task, llm=model, tools=tools, browser_profile=browser_profile) history = await agent.run() result = history.final_result() if result: parsed: PersonList = PersonList.model_validate_json(result) for person in parsed.people: print(f'{person.name} - {person.email}') else: print('No result') if __name__ == '__main__': asyncio.run(main())
python
MIT
630f85dd05127c9d42810a5db235a14f5bac9043
2026-01-04T14:38:16.467592Z
false
browser-use/browser-use
https://github.com/browser-use/browser-use/blob/630f85dd05127c9d42810a5db235a14f5bac9043/examples/custom-functions/notification.py
examples/custom-functions/notification.py
import asyncio import os import sys sys.path.append(os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))) from dotenv import load_dotenv load_dotenv() from browser_use import ActionResult, Agent, ChatOpenAI, Tools tools = Tools() @tools.registry.action('Done with task') async def done(text: str): import yagmail # type: ignore # To send emails use # STEP 1: go to https://support.google.com/accounts/answer/185833 # STEP 2: Create an app password (you can't use here your normal gmail password) # STEP 3: Use the app password in the code below for the password yag = yagmail.SMTP('your_email@gmail.com', 'your_app_password') yag.send( to='recipient@example.com', subject='Test Email', contents=f'result\n: {text}', ) return ActionResult(is_done=True, extracted_content='Email sent!') async def main(): task = 'go to brower-use.com and then done' model = ChatOpenAI(model='gpt-4.1-mini') agent = Agent(task=task, llm=model, tools=tools) await agent.run() if __name__ == '__main__': asyncio.run(main())
python
MIT
630f85dd05127c9d42810a5db235a14f5bac9043
2026-01-04T14:38:16.467592Z
false
browser-use/browser-use
https://github.com/browser-use/browser-use/blob/630f85dd05127c9d42810a5db235a14f5bac9043/examples/custom-functions/cua.py
examples/custom-functions/cua.py
""" OpenAI Computer Use Assistant (CUA) Integration This example demonstrates how to integrate OpenAI's Computer Use Assistant as a fallback action when standard browser actions are insufficient to achieve the desired goal. The CUA can perform complex computer interactions that might be difficult to achieve through regular browser-use actions. """ import asyncio import base64 import os import sys from io import BytesIO from PIL import Image sys.path.append(os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))) from dotenv import load_dotenv load_dotenv() from openai import AsyncOpenAI from pydantic import BaseModel, Field from browser_use import Agent, ChatOpenAI, Tools from browser_use.agent.views import ActionResult from browser_use.browser import BrowserSession class OpenAICUAAction(BaseModel): """Parameters for OpenAI Computer Use Assistant action.""" description: str = Field(..., description='Description of your next goal') async def handle_model_action(browser_session: BrowserSession, action) -> ActionResult: """ Given a computer action (e.g., click, double_click, scroll, etc.), execute the corresponding operation using CDP. """ action_type = action.type ERROR_MSG: str = 'Could not execute the CUA action.' if not browser_session.agent_focus_target_id: return ActionResult(error='No active browser session') # Get CDP session for the focused target using the public API try: cdp_session = await browser_session.get_or_create_cdp_session(browser_session.agent_focus_target_id, focus=False) except Exception as e: return ActionResult(error=f'Failed to get CDP session: {e}') try: match action_type: case 'click': x, y = action.x, action.y button = action.button print(f"Action: click at ({x}, {y}) with button '{button}'") # Not handling things like middle click, etc. if button != 'left' and button != 'right': button = 'left' # Use CDP to click await browser_session.cdp_client.send.Input.dispatchMouseEvent( params={ 'type': 'mousePressed', 'x': x, 'y': y, 'button': button, 'clickCount': 1, }, session_id=cdp_session.session_id, ) await browser_session.cdp_client.send.Input.dispatchMouseEvent( params={ 'type': 'mouseReleased', 'x': x, 'y': y, 'button': button, }, session_id=cdp_session.session_id, ) msg = f'Clicked at ({x}, {y}) with button {button}' return ActionResult(extracted_content=msg, include_in_memory=True, long_term_memory=msg) case 'scroll': x, y = action.x, action.y scroll_x, scroll_y = action.scroll_x, action.scroll_y print(f'Action: scroll at ({x}, {y}) with offsets (scroll_x={scroll_x}, scroll_y={scroll_y})') # Move mouse to position first await browser_session.cdp_client.send.Input.dispatchMouseEvent( params={ 'type': 'mouseMoved', 'x': x, 'y': y, }, session_id=cdp_session.session_id, ) # Execute scroll using JavaScript await browser_session.cdp_client.send.Runtime.evaluate( params={ 'expression': f'window.scrollBy({scroll_x}, {scroll_y})', }, session_id=cdp_session.session_id, ) msg = f'Scrolled at ({x}, {y}) with offsets (scroll_x={scroll_x}, scroll_y={scroll_y})' return ActionResult(extracted_content=msg, include_in_memory=True, long_term_memory=msg) case 'keypress': keys = action.keys for k in keys: print(f"Action: keypress '{k}'") # A simple mapping for common keys; expand as needed. key_code = k if k.lower() == 'enter': key_code = 'Enter' elif k.lower() == 'space': key_code = 'Space' # Use CDP to send key await browser_session.cdp_client.send.Input.dispatchKeyEvent( params={ 'type': 'keyDown', 'key': key_code, }, session_id=cdp_session.session_id, ) await browser_session.cdp_client.send.Input.dispatchKeyEvent( params={ 'type': 'keyUp', 'key': key_code, }, session_id=cdp_session.session_id, ) msg = f'Pressed keys: {keys}' return ActionResult(extracted_content=msg, include_in_memory=True, long_term_memory=msg) case 'type': text = action.text print(f'Action: type text: {text}') # Type text character by character for char in text: await browser_session.cdp_client.send.Input.dispatchKeyEvent( params={ 'type': 'char', 'text': char, }, session_id=cdp_session.session_id, ) msg = f'Typed text: {text}' return ActionResult(extracted_content=msg, include_in_memory=True, long_term_memory=msg) case 'wait': print('Action: wait') await asyncio.sleep(2) msg = 'Waited for 2 seconds' return ActionResult(extracted_content=msg, include_in_memory=True, long_term_memory=msg) case 'screenshot': # Nothing to do as screenshot is taken at each turn print('Action: screenshot') return ActionResult(error=ERROR_MSG) # Handle other actions here case _: print(f'Unrecognized action: {action}') return ActionResult(error=ERROR_MSG) except Exception as e: print(f'Error handling action {action}: {e}') return ActionResult(error=ERROR_MSG) tools = Tools() @tools.registry.action( 'Use OpenAI Computer Use Assistant (CUA) as a fallback when standard browser actions cannot achieve the desired goal. This action sends a screenshot and description to OpenAI CUA and executes the returned computer use actions.', param_model=OpenAICUAAction, ) async def openai_cua_fallback(params: OpenAICUAAction, browser_session: BrowserSession): """ Fallback action that uses OpenAI's Computer Use Assistant to perform complex computer interactions when standard browser actions are insufficient. """ print(f'🎯 CUA Action Starting - Goal: {params.description}') try: # Get browser state summary state = await browser_session.get_browser_state_summary() page_info = state.page_info if not page_info: raise Exception('Page info not found - cannot execute CUA action') print(f'πŸ“ Viewport size: {page_info.viewport_width}x{page_info.viewport_height}') screenshot_b64 = state.screenshot if not screenshot_b64: raise Exception('Screenshot not found - cannot execute CUA action') print(f'πŸ“Έ Screenshot captured (base64 length: {len(screenshot_b64)} chars)') # Debug: Check screenshot dimensions image = Image.open(BytesIO(base64.b64decode(screenshot_b64))) print(f'πŸ“ Screenshot actual dimensions: {image.size[0]}x{image.size[1]}') # rescale the screenshot to the viewport size image = image.resize((page_info.viewport_width, page_info.viewport_height)) # Save as PNG to bytes buffer buffer = BytesIO() image.save(buffer, format='PNG') buffer.seek(0) # Convert to base64 screenshot_b64 = base64.b64encode(buffer.getvalue()).decode('utf-8') print(f'πŸ“Έ Rescaled screenshot to viewport size: {page_info.viewport_width}x{page_info.viewport_height}') client = AsyncOpenAI(api_key=os.getenv('OPENAI_API_KEY')) print('πŸ”„ Sending request to OpenAI CUA...') prompt = f""" You will be given an action to execute and screenshot of the current screen. Output one computer_call object that will achieve this goal. Goal: {params.description} """ response = await client.responses.create( model='computer-use-preview', tools=[ { 'type': 'computer_use_preview', 'display_width': page_info.viewport_width, 'display_height': page_info.viewport_height, 'environment': 'browser', } ], input=[ { 'role': 'user', 'content': [ {'type': 'input_text', 'text': prompt}, { 'type': 'input_image', 'detail': 'auto', 'image_url': f'data:image/png;base64,{screenshot_b64}', }, ], } ], truncation='auto', temperature=0.1, ) print(f'πŸ“₯ CUA response received: {response}') computer_calls = [item for item in response.output if item.type == 'computer_call'] computer_call = computer_calls[0] if computer_calls else None if not computer_call: raise Exception('No computer calls found in CUA response') action = computer_call.action print(f'🎬 Executing CUA action: {action.type} - {action}') action_result = await handle_model_action(browser_session, action) await asyncio.sleep(0.1) print('βœ… CUA action completed successfully') return action_result except Exception as e: msg = f'Error executing CUA action: {e}' print(f'❌ {msg}') return ActionResult(error=msg) async def main(): # Initialize the language model llm = ChatOpenAI( model='o4-mini', temperature=1.0, ) # Create browser session browser_session = BrowserSession() # Example task that might require CUA fallback # This could be a complex interaction that's difficult with standard actions task = """ Go to https://csreis.github.io/tests/cross-site-iframe.html Click on "Go cross-site, complex page" using index Use the OpenAI CUA fallback to click on "Tree is open..." link. """ # Create agent with our custom tools that includes CUA fallback agent = Agent( task=task, llm=llm, tools=tools, browser_session=browser_session, ) print('πŸš€ Starting agent with CUA fallback support...') print(f'Task: {task}') print('-' * 50) try: # Run the agent result = await agent.run() print(f'\nβœ… Task completed! Result: {result}') except Exception as e: print(f'\n❌ Error running agent: {e}') finally: # Clean up browser session await browser_session.kill() print('\n🧹 Browser session closed') if __name__ == '__main__': # Example of different scenarios where CUA might be useful print('πŸ”§ OpenAI Computer Use Assistant (CUA) Integration Example') print('=' * 60) print() print("This example shows how to integrate OpenAI's CUA as a fallback action") print('when standard browser-use actions cannot achieve the desired goal.') print() print('CUA is particularly useful for:') print('β€’ Complex mouse interactions (drag & drop, precise clicking)') print('β€’ Keyboard shortcuts and key combinations') print('β€’ Actions that require pixel-perfect precision') print("β€’ Custom UI elements that don't respond to standard actions") print() print('Make sure you have OPENAI_API_KEY set in your environment!') print() # Check if OpenAI API key is available if not os.getenv('OPENAI_API_KEY'): print('❌ Error: OPENAI_API_KEY environment variable not set') print('Please set your OpenAI API key to use CUA integration') sys.exit(1) # Run the example asyncio.run(main())
python
MIT
630f85dd05127c9d42810a5db235a14f5bac9043
2026-01-04T14:38:16.467592Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/scripts/qwen_omni_merge.py
scripts/qwen_omni_merge.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Why we need this script for qwen_omni? Because the qwen_omni model is constructed by two parts: 1. [Thinker]:[audio_encoder, vision_encoder, LLM backbone], which our repository does support to post-training. 2. [Talker]: [audio_decoder, wave_model], which is not supported to post-training without specific tokenizer. When we post-training the model, we exactly train the [Thinker] part, and the [Talker] part is dropped. So, to get the complete model, we need to merge the [Talker] part back to the [Thinker] part. LoRA mode: [Thinker + LoRA weights] + [Original Talker] -> [Omni model] Full mode: [Thinker] + [Original Talker] -> [Omni model] For Processor, we do saved the processor from trained model instead of the original model. """ import os import shutil import fire from peft import PeftModel from transformers import AutoConfig, AutoModelForTextToWaveform, AutoProcessor from transformers.utils import cached_file def merge_lora( model_path: str, lora_path: str, save_path: str = "./merged_model_checkpoint", extra_file: str = "spk_dict.pt", submodule_name: str = "thinker", ): """Load the original model, merge the LoRA weights. For a specified submodule, and save the final merged model along with its configurations. Args: model_path (str): Path to the original model directory. lora_path (str): Path to the directory containing LoRA weights. save_path (str): Directory where the merged model and configurations will be saved. extra_file (str): Name of the extra file to be copied (default: "spk_dict.pt"). submodule_name (str): Name of the submodule to merge (default: "thinker"). """ # 1. Load the original model model = AutoModelForTextToWaveform.from_pretrained(model_path, torch_dtype="auto", device_map="cpu") print("Successfully loaded the original model.") # 2. Extract the submodule to be merged (e.g., model.thinker) if not hasattr(model, submodule_name): raise AttributeError(f"The model does not have a submodule named '{submodule_name}'.") base_submodule = getattr(model, submodule_name) print(f"Successfully extracted submodule: {submodule_name}.") # 3. Load the LoRA weights onto the extracted submodule lora_model = PeftModel.from_pretrained(base_submodule, lora_path) processor = AutoProcessor.from_pretrained(lora_path) print("Successfully loaded LoRA weights and processor.") # 4. Merge the LoRA weights into the submodule and unload the LoRA modules merged_submodule = lora_model.merge_and_unload() print("Successfully merged LoRA weights.") # 5. Replace the original submodule with the merged submodule in the model setattr(model, submodule_name, merged_submodule) # 6. Save the final merged model along with the tokenizer and processor configuration model.save_pretrained(save_path) processor.save_pretrained(save_path) print(f"Merged model and processor saved to {save_path}.") try: source_file = cached_file(path_or_repo_id=model_path, filename=extra_file) shutil.copy(source_file, os.path.join(save_path, extra_file)) print(f"File '{extra_file}' copied from {model_path} to {save_path}.") except Exception: print(f"File '{extra_file}' not found in {model_path}, skipping copy.") def save_full_model( model_path: str, thinker_path: str, save_path: str = "./merged_model_checkpoint", extra_file: str = "spk_dict.pt", ): """Load the saved thinker module and the original model, replace the thinker in the original model. Then save the complete model along with its tokenizer and processor configuration. Args: model_path (str): Directory path of the original model. thinker_path (str): Path to the saved thinker weights. save_path (str): Directory where the merged model and configurations will be saved. extra_file (str): Name of the extra file to be copied (default: "spk_dict.pt"). """ # 1. Load the saved thinker module and the original model config = AutoConfig.from_pretrained(model_path) if getattr(config, "model_type") == "qwen2_5_omni": from transformers.models.qwen2_5_omni import Qwen2_5OmniThinkerForConditionalGeneration # type: ignore ThinkerClass = Qwen2_5OmniThinkerForConditionalGeneration elif getattr(config, "model_type") == "qwen3_omni_moe": from transformers.models.qwen3_omni_moe import Qwen3OmniMoeThinkerForConditionalGeneration # type: ignore ThinkerClass = Qwen3OmniMoeThinkerForConditionalGeneration else: raise ValueError(f"Unsupported model type: {getattr(config, 'model_type')}.") thinker = ThinkerClass.from_pretrained(thinker_path, torch_dtype="auto", device_map="cpu") base_model = AutoModelForTextToWaveform.from_pretrained(model_path, torch_dtype="auto", device_map="cpu") base_model.thinker = thinker processor = AutoProcessor.from_pretrained(thinker_path) print("Successfully loaded model weights and processor.") # 2. Save the complete model along with its tokenizer and processor configuration base_model.save_pretrained(save_path) processor.save_pretrained(save_path) print(f"Merged model and processor saved to {save_path}.") # 3. Copy the extra file from the base model directory to the save_path try: source_file = cached_file(path_or_repo_id=model_path, filename=extra_file) shutil.copy(source_file, os.path.join(save_path, extra_file)) print(f"File '{extra_file}' copied from {model_path} to {save_path}.") except Exception: print(f"File '{extra_file}' not found in {model_path}, skipping copy.") if __name__ == "__main__": fire.Fire({"save_full": save_full_model, "merge_lora": merge_lora})
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/scripts/bench_qwen.py
scripts/bench_qwen.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import os from dataclasses import dataclass from typing import Any import fire import torch from peft import PeftModel from torch.utils.data import Dataset from transformers import DataCollatorForSeq2Seq, Qwen2_5_VLProcessor from llamafactory.extras.constants import IGNORE_INDEX from llamafactory.hparams import get_train_args from llamafactory.model import load_model, load_tokenizer from llamafactory.train.callbacks import LogCallback from llamafactory.train.sft.trainer import CustomSeq2SeqTrainer class DummyDataset(Dataset): def __init__(self, size: int = 1000, seq_length: int = 1024, processor: Qwen2_5_VLProcessor = None): self.size = size self.seq_length = seq_length self.vocab_size = 32768 self.processor = processor image_token_num = 18 * 18 // (2 * 2) image_t = 2 self.text_seqlen = seq_length // 4 # 25% text video_seq_length = self.seq_length - self.text_seqlen - image_t * image_token_num video_t = video_seq_length // image_token_num self.image_size = [18 * 18 * image_t, 1176] self.image_grid_thw = torch.tensor([[1, 18, 18]] * image_t, dtype=torch.long) self.image_seqlen = image_t * image_token_num self.video_size = [18 * 18 * video_t, 1176] self.video_grid_thw = torch.tensor([[video_t, 18, 18]], dtype=torch.long) self.video_seqlen = video_t * image_token_num def __len__(self): return self.size def __getitem__(self, index: int): input_ids = torch.randint(low=0, high=self.vocab_size, size=(self.seq_length,)) input_ids[: self.image_seqlen] = self.processor.image_token_id input_ids[self.image_seqlen : self.image_seqlen + self.video_seqlen] = self.processor.video_token_id attention_mask = torch.ones((self.seq_length,), dtype=torch.long) labels = input_ids.clone() labels[: self.image_seqlen + self.video_seqlen] = IGNORE_INDEX pixel_values = torch.rand(self.image_size, dtype=torch.float32) pixel_values_videos = torch.rand(self.video_size, dtype=torch.float32) return { "input_ids": input_ids, "attention_mask": attention_mask, "labels": labels, "pixel_values": pixel_values, "pixel_values_videos": pixel_values_videos, "image_grid_thw": self.image_grid_thw, "video_grid_thw": self.video_grid_thw, } @dataclass class MultiModalDataCollatorForSeq2Seq(DataCollatorForSeq2Seq): def __post_init__(self): if isinstance(self.model, PeftModel): self.model = self.model.base_model.model if self.model is not None and hasattr(self.model, "get_rope_index"): # for qwen2vl mrope self.get_rope_func = self.model.get_rope_index # transformers < 4.52.0 or qwen2.5 omni elif self.model is not None and hasattr(self.model, "model") and hasattr(self.model.model, "get_rope_index"): self.get_rope_func = self.model.model.get_rope_index # transformers >= 4.52.0 else: self.get_rope_func = None def __call__(self, features: list[dict[str, Any]]) -> dict[str, "torch.Tensor"]: batch_pixel_values = [feature.pop("pixel_values") for feature in features] batch_pixel_values_videos = [feature.pop("pixel_values_videos") for feature in features] batch_image_grid_thw = [feature.pop("image_grid_thw") for feature in features] batch_video_grid_thw = [feature.pop("video_grid_thw") for feature in features] batch: dict[str, torch.Tensor] = super().__call__(features) batch["pixel_values"] = torch.cat(batch_pixel_values, dim=0) batch["pixel_values_videos"] = torch.cat(batch_pixel_values_videos, dim=0) batch["image_grid_thw"] = torch.cat(batch_image_grid_thw, dim=0) batch["video_grid_thw"] = torch.cat(batch_video_grid_thw, dim=0) if self.get_rope_func is not None: rope_index_kwargs = { "input_ids": batch["input_ids"], "image_grid_thw": batch["image_grid_thw"], "video_grid_thw": batch["video_grid_thw"], "attention_mask": (batch["attention_mask"] >= 1).float(), } batch["position_ids"], batch["rope_deltas"] = self.get_rope_func(**rope_index_kwargs) if "position_ids" not in batch or batch["position_ids"].dim() != 3: raise ValueError("Qwen2VL requires 3D position ids for mrope.") return batch def bench_qwen( model_name_or_path: str = "Qwen/Qwen2-VL-7B-Instruct", batch_size: int = 1, seq_length: int = 2048, liger_kernel: bool = False, deepspeed_stage: int = 3, ): os.environ["LLAMABOARD_ENABLED"] = "true" os.environ["LLAMABOARD_WORKDIR"] = "output/dummy_dir" args = { "model_name_or_path": model_name_or_path, "enable_liger_kernel": liger_kernel, "stage": "sft", "do_train": True, "finetuning_type": "full", "dataset": "alpaca_en_demo", "template": "qwen2_vl", "cutoff_len": seq_length, "output_dir": "output/dummy_dir", "logging_steps": 10, "save_strategy": "no", "save_only_model": True, "overwrite_output_dir": True, "per_device_train_batch_size": batch_size, "max_steps": 1000, "bf16": True, "include_num_input_tokens_seen": True, "report_to": "none", } if deepspeed_stage in [2, 3]: args["deepspeed"] = f"examples/deepspeed/ds_z{deepspeed_stage}_config.json" model_args, _, training_args, finetuning_args, _ = get_train_args(args) tokenizer_module = load_tokenizer(model_args) tokenizer = tokenizer_module["tokenizer"] trainset = DummyDataset(size=100000, seq_length=seq_length, processor=tokenizer_module["processor"]) model = load_model(tokenizer, model_args, finetuning_args, training_args.do_train) data_collator = MultiModalDataCollatorForSeq2Seq( tokenizer=tokenizer, model=model, pad_to_multiple_of=8, label_pad_token_id=IGNORE_INDEX ) trainer = CustomSeq2SeqTrainer( model=model, args=training_args, finetuning_args=finetuning_args, data_collator=data_collator, callbacks=[LogCallback()], train_dataset=trainset, **tokenizer_module, ) trainer.train(resume_from_checkpoint=training_args.resume_from_checkpoint) if __name__ == "__main__": fire.Fire(bench_qwen)
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/scripts/pissa_init.py
scripts/pissa_init.py
# Copyright 2025 HuggingFace Inc. and the LlamaFactory team. # # This code is based on the HuggingFace's PEFT library. # https://github.com/huggingface/peft/blob/v0.11.0/examples/pissa_finetuning/preprocess.py # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import os from typing import TYPE_CHECKING import fire from peft import LoraConfig, TaskType, get_peft_model from transformers import AutoModelForCausalLM, AutoTokenizer if TYPE_CHECKING: from transformers import PreTrainedModel def quantize_pissa( model_name_or_path: str, output_dir: str, pissa_iter: int = 16, lora_alpha: int = None, lora_rank: int = 16, lora_dropout: float = 0, lora_target: tuple = ("q_proj", "v_proj"), save_safetensors: bool = True, ): r"""Initialize LoRA weights with Principal Singular values and Singular vectors Adaptation (PiSSA). Usage: python pissa_init.py --model_name_or_path path_to_model --output_dir output_dir """ if isinstance(lora_target, str): lora_target = [name.strip() for name in lora_target.split(",")] tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, trust_remote_code=True) model = AutoModelForCausalLM.from_pretrained(model_name_or_path, trust_remote_code=True, torch_dtype="auto") lora_config = LoraConfig( task_type=TaskType.CAUSAL_LM, r=lora_rank, lora_alpha=lora_alpha if lora_alpha is not None else lora_rank * 2, lora_dropout=lora_dropout, target_modules=lora_target, init_lora_weights="pissa" if pissa_iter == -1 else f"pissa_niter_{pissa_iter}", ) # Init PiSSA model peft_model = get_peft_model(model, lora_config) pissa_dir = os.path.join(output_dir, "pissa_init") # Save PiSSA model setattr(peft_model.peft_config["default"], "base_model_name_or_path", os.path.abspath(output_dir)) setattr(peft_model.peft_config["default"], "init_lora_weights", True) # don't apply pissa again peft_model.save_pretrained(pissa_dir, safe_serialization=save_safetensors) print(f"Adapter weights saved in {pissa_dir}") # Save base model base_model: PreTrainedModel = peft_model.unload() base_model.save_pretrained(output_dir, safe_serialization=save_safetensors) tokenizer.save_pretrained(output_dir) print(f"Model weights saved in {output_dir}") print("- Fine-tune this model with:") print(f"model_name_or_path: {output_dir}") print(f"adapter_name_or_path: {pissa_dir}") print("finetuning_type: lora") print("pissa_init: false") print("pissa_convert: true") print("- and optionally with:") print("quantization_bit: 4") if __name__ == "__main__": fire.Fire(quantize_pissa)
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/scripts/eval_bleu_rouge.py
scripts/eval_bleu_rouge.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import json import logging import time import fire from datasets import load_dataset try: import jieba # type: ignore from nltk.translate.bleu_score import SmoothingFunction, sentence_bleu # type: ignore from rouge_chinese import Rouge # type: ignore jieba.setLogLevel(logging.CRITICAL) jieba.initialize() except ImportError: print("Please install llamafactory with `pip install -r requirements/metrics.txt`.") raise def compute_metrics(sample): hypothesis = list(jieba.cut(sample["predict"])) reference = list(jieba.cut(sample["label"])) bleu_score = sentence_bleu( [list(sample["label"])], list(sample["predict"]), smoothing_function=SmoothingFunction().method3, ) if len(" ".join(hypothesis).split()) == 0 or len(" ".join(reference).split()) == 0: result = {"rouge-1": {"f": 0.0}, "rouge-2": {"f": 0.0}, "rouge-l": {"f": 0.0}} else: rouge = Rouge() scores = rouge.get_scores(" ".join(hypothesis), " ".join(reference)) result = scores[0] metric_result = {} for k, v in result.items(): metric_result[k] = round(v["f"] * 100, 4) metric_result["bleu-4"] = round(bleu_score * 100, 4) return metric_result def main(filename: str): start_time = time.time() dataset = load_dataset("json", data_files=filename, split="train") dataset = dataset.map(compute_metrics, num_proc=8, remove_columns=dataset.column_names) score_dict = dataset.to_dict() average_score = {} for task, scores in sorted(score_dict.items(), key=lambda x: x[0]): print(f"{task}: {sum(scores) / len(scores):.4f}") average_score[task] = sum(scores) / len(scores) with open("predictions_score.json", "w", encoding="utf-8") as f: json.dump(average_score, f, indent=4) print(f"\nDone in {time.time() - start_time:.3f}s.\nScore file saved to predictions_score.json") if __name__ == "__main__": fire.Fire(main)
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/scripts/llama_pro.py
scripts/llama_pro.py
# Copyright 2025 Tencent Inc. and the LlamaFactory team. # # This code is inspired by the Tencent's LLaMA-Pro library. # https://github.com/TencentARC/LLaMA-Pro/blob/main/scripts/block_expansion.py # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import json import os from collections import OrderedDict from typing import TYPE_CHECKING import fire import torch from huggingface_hub import split_torch_state_dict_into_shards from safetensors.torch import save_file from tqdm import tqdm from transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer, PreTrainedModel from transformers.modeling_utils import SAFE_WEIGHTS_INDEX_NAME, SAFE_WEIGHTS_NAME, WEIGHTS_INDEX_NAME, WEIGHTS_NAME if TYPE_CHECKING: from transformers import PretrainedConfig def change_name(name: str, old_index: int, new_index: int) -> str: return name.replace(f".{old_index:d}.", f".{new_index:d}.") def block_expansion( model_name_or_path: str, output_dir: str, num_expand: int, shard_size: str = "5GB", save_safetensors: bool = True, ): r"""Perform block expansion for LLaMA, Mistral, Qwen2 or Yi models. Usage: python llama_pro.py --model_name_or_path meta-llama/Llama-2-7b-hf --output_dir llama2_pro --num_expand 8 """ config: PretrainedConfig = AutoConfig.from_pretrained(model_name_or_path, trust_remote_code=True) num_layers = getattr(config, "num_hidden_layers") if num_layers % num_expand != 0: raise ValueError(f"`num_layers` {num_layers} should be divisible by `num_expand` {num_expand}.") setattr(config, "num_hidden_layers", num_layers + num_expand) config.save_pretrained(output_dir) tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, trust_remote_code=True) tokenizer.save_pretrained(output_dir) print(f"Expanding model of {num_layers} layers to {num_layers + num_expand} layers.") model = AutoModelForCausalLM.from_pretrained( model_name_or_path, torch_dtype="auto", device_map="cpu", trust_remote_code=True, low_cpu_mem_usage=True ) assert isinstance(model, PreTrainedModel) # type hint if save_safetensors and getattr(model.config, "tie_word_embeddings", False): del model.lm_head # safetensors does not allow shared weights split = num_layers // num_expand layer_cnt = 0 state_dict = model.state_dict() output_state_dict: dict[str, torch.Tensor] = OrderedDict() for i in range(num_layers): for key, value in state_dict.items(): if f".{i:d}." in key: output_state_dict[change_name(key, i, layer_cnt)] = value print(f"Add layer {layer_cnt} copied from layer {i}.") layer_cnt += 1 if (i + 1) % split == 0: for key, value in state_dict.items(): if f".{i:d}." in key: if "down_proj" in key or "o_proj" in key: output_state_dict[change_name(key, i, layer_cnt)] = torch.zeros_like(value) else: output_state_dict[change_name(key, i, layer_cnt)] = torch.clone(value) print(f"Add layer {layer_cnt} expanded from layer {i}.") layer_cnt += 1 for key, value in state_dict.items(): if key not in output_state_dict: output_state_dict[key] = value weights_name = SAFE_WEIGHTS_NAME if save_safetensors else WEIGHTS_NAME filename_pattern = weights_name.replace(".bin", "{suffix}.bin").replace(".safetensors", "{suffix}.safetensors") state_dict_split = split_torch_state_dict_into_shards( output_state_dict, filename_pattern=filename_pattern, max_shard_size=shard_size ) for shard_file, tensors in tqdm(state_dict_split.filename_to_tensors.items(), desc="Save weights"): shard = {tensor: output_state_dict[tensor].contiguous() for tensor in tensors} if save_safetensors: save_file(shard, os.path.join(output_dir, shard_file), metadata={"format": "pt"}) else: torch.save(shard, os.path.join(output_dir, shard_file)) if not state_dict_split.is_sharded: print(f"Model weights saved in {os.path.join(output_dir, weights_name)}.") else: index = { "metadata": state_dict_split.metadata, "weight_map": state_dict_split.tensor_to_filename, } index_name = SAFE_WEIGHTS_INDEX_NAME if save_safetensors else WEIGHTS_INDEX_NAME with open(os.path.join(output_dir, index_name), "w", encoding="utf-8") as f: json.dump(index, f, indent=2, sort_keys=True) print(f"Model weights saved in {output_dir}.") print("- Fine-tune this model with:") print(f"model_name_or_path: {output_dir}") print("finetuning_type: freeze") print(f"freeze_trainable_layers: {num_expand}") print("use_llama_pro: true") if __name__ == "__main__": fire.Fire(block_expansion)
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/scripts/megatron_merge.py
scripts/megatron_merge.py
# Copyright 2025 the ROLL team and the LlamaFactory team. # # This code is modified from the ROLL library. # https://github.com/alibaba/ROLL/blob/main/mcore_adapter/tools/convert.py # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import os import fire import torch from mcore_adapter.models.converter.post_converter import convert_checkpoint_to_hf, convert_checkpoint_to_mca from mcore_adapter.training_args import DistributingParallelArguments from mcore_adapter.utils import get_logger from transformers import AutoConfig logger = get_logger(__name__) def convert_mca_to_hf( checkpoint_path: str, output_path: str = "./output", bf16: bool = False, fp16: bool = False, convert_model_max_length: int | None = None, ): """Convert megatron checkpoint to HuggingFace format. Args: checkpoint_path: Path to the checkpoint to convert output_path: Path to save the converted checkpoint bf16: Use bfloat16 precision fp16: Use float16 precision convert_model_max_length: Change the model_max_length in hf config.json """ if bf16 and fp16: raise ValueError("bf16 and fp16 cannot be both True.") torch_dtype = None if bf16: torch_dtype = torch.bfloat16 elif fp16: torch_dtype = torch.float16 convert_checkpoint_to_hf(checkpoint_path, output_path, torch_dtype=torch_dtype) if convert_model_max_length is not None: config = AutoConfig.from_pretrained(output_path, trust_remote_code=True) config.model_max_length = convert_model_max_length config.save_pretrained(output_path) def convert( checkpoint_path: str, output_path: str = "./output", bf16: bool = False, fp16: bool = False, convert_model_max_length: int | None = None, tensor_model_parallel_size: int = 1, pipeline_model_parallel_size: int = 1, expert_model_parallel_size: int = 1, virtual_pipeline_model_parallel_size: int | None = None, ): """Convert checkpoint between MCA and HuggingFace formats. Args: checkpoint_path: Path to the checkpoint to convert output_path: Path to save the converted checkpoint bf16: Use bfloat16 precision fp16: Use float16 precision convert_model_max_length: Change the model_max_length in hf config.json tensor_model_parallel_size: Tensor model parallel size pipeline_model_parallel_size: Pipeline model parallel size expert_model_parallel_size: Expert model parallel size virtual_pipeline_model_parallel_size: Virtual pipeline model parallel size """ if bf16 and fp16: raise ValueError("bf16 and fp16 cannot be both True.") mca_config_path = os.path.join(checkpoint_path, "mca_config.json") from_mca = os.path.exists(mca_config_path) if not from_mca: dist_args = DistributingParallelArguments( tensor_model_parallel_size=tensor_model_parallel_size, pipeline_model_parallel_size=pipeline_model_parallel_size, expert_model_parallel_size=expert_model_parallel_size, virtual_pipeline_model_parallel_size=virtual_pipeline_model_parallel_size, ) convert_checkpoint_to_mca( checkpoint_path, output_path, dist_args, bf16=bf16, fp16=fp16, ) else: convert_mca_to_hf( checkpoint_path=checkpoint_path, output_path=output_path, bf16=bf16, fp16=fp16, convert_model_max_length=convert_model_max_length, ) def main(): fire.Fire(convert) if __name__ == "__main__": main()
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/scripts/vllm_infer.py
scripts/vllm_infer.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import gc import json import time import av import fire from datasets import load_dataset from eval_bleu_rouge import compute_metrics from tqdm import tqdm from transformers import Seq2SeqTrainingArguments from llamafactory.data import get_dataset, get_template_and_fix_tokenizer from llamafactory.extras.constants import IGNORE_INDEX from llamafactory.extras.misc import get_device_count from llamafactory.extras.packages import is_vllm_available from llamafactory.hparams import get_infer_args from llamafactory.model import load_tokenizer if is_vllm_available(): from vllm import LLM, SamplingParams from vllm.lora.request import LoRARequest def _need_video_kwargs(template): NEEDED_TEMPLATE = ["qwen3_vl", "glm4v"] if any(t in template for t in NEEDED_TEMPLATE): return True return False def vllm_infer( model_name_or_path: str, adapter_name_or_path: str = None, dataset: str = "alpaca_en_demo", dataset_dir: str = "data", template: str = "default", cutoff_len: int = 2048, max_samples: int | None = None, vllm_config: str = "{}", save_name: str = "generated_predictions.jsonl", matrix_save_name: str = None, temperature: float = 0.95, top_p: float = 0.7, top_k: int = 50, max_new_tokens: int = 1024, repetition_penalty: float = 1.0, skip_special_tokens: bool = True, default_system: str | None = None, enable_thinking: bool = True, seed: int | None = None, pipeline_parallel_size: int = 1, image_max_pixels: int = 768 * 768, image_min_pixels: int = 32 * 32, video_fps: float = 2.0, video_maxlen: int = 128, batch_size: int = 1024, ): r"""Perform batch generation using vLLM engine, which supports tensor parallelism. Usage: python vllm_infer.py --model_name_or_path meta-llama/Llama-2-7b-hf --template llama --dataset alpaca_en_demo """ if pipeline_parallel_size > get_device_count(): raise ValueError("Pipeline parallel size should be smaller than the number of gpus.") model_args, data_args, _, generating_args = get_infer_args( dict( model_name_or_path=model_name_or_path, adapter_name_or_path=adapter_name_or_path, dataset=dataset, dataset_dir=dataset_dir, template=template, cutoff_len=cutoff_len, max_samples=max_samples, preprocessing_num_workers=16, default_system=default_system, enable_thinking=enable_thinking, vllm_config=vllm_config, temperature=temperature, top_p=top_p, top_k=top_k, max_new_tokens=max_new_tokens, repetition_penalty=repetition_penalty, ) ) training_args = Seq2SeqTrainingArguments(output_dir="dummy_dir") tokenizer_module = load_tokenizer(model_args) tokenizer = tokenizer_module["tokenizer"] template_obj = get_template_and_fix_tokenizer(tokenizer, data_args) template_obj.mm_plugin.expand_mm_tokens = False # for vllm generate engine_args = { "model": model_args.model_name_or_path, "trust_remote_code": True, "dtype": model_args.infer_dtype, "max_model_len": cutoff_len + max_new_tokens, "tensor_parallel_size": (get_device_count() // pipeline_parallel_size) or 1, "pipeline_parallel_size": pipeline_parallel_size, "disable_log_stats": True, "enable_lora": model_args.adapter_name_or_path is not None, } if template_obj.mm_plugin.__class__.__name__ != "BasePlugin": engine_args["limit_mm_per_prompt"] = {"image": 4, "video": 2, "audio": 2} if isinstance(model_args.vllm_config, dict): engine_args.update(model_args.vllm_config) model_preparation_start_time = time.time() llm = LLM(**engine_args) # load datasets dataset_module = get_dataset(template_obj, model_args, data_args, training_args, "ppo", **tokenizer_module) train_dataset = dataset_module["train_dataset"] sampling_params = SamplingParams( repetition_penalty=generating_args.repetition_penalty or 1.0, # repetition_penalty must > 0 temperature=generating_args.temperature, top_p=generating_args.top_p or 1.0, # top_p must > 0 top_k=generating_args.top_k or -1, # top_k must > 0 stop_token_ids=template_obj.get_stop_token_ids(tokenizer), max_tokens=generating_args.max_new_tokens, skip_special_tokens=skip_special_tokens, seed=seed, ) if model_args.adapter_name_or_path is not None: lora_request = LoRARequest("default", 1, model_args.adapter_name_or_path[0]) else: lora_request = None # Store all results in these lists all_prompts, all_preds, all_labels = [], [], [] need_video_kwargs = _need_video_kwargs(template) model_predict_start_time = time.time() # Add batch process to avoid the issue of too many files opened for i in tqdm(range(0, len(train_dataset), batch_size), desc="Processing batched inference"): vllm_inputs, prompts, labels = [], [], [] batch = train_dataset[i : min(i + batch_size, len(train_dataset))] for j in range(len(batch["input_ids"])): if batch["images"][j] is not None: image = batch["images"][j] multi_modal_data = { "image": template_obj.mm_plugin._regularize_images( image, image_max_pixels=image_max_pixels, image_min_pixels=image_min_pixels )["images"] } elif batch["videos"][j] is not None: video_metadata, video_metadata_kwargs = None, None video = batch["videos"][j] multi_modal_data = { "video": template_obj.mm_plugin._regularize_videos( video, image_max_pixels=image_max_pixels, image_min_pixels=image_min_pixels, video_fps=video_fps, video_maxlen=video_maxlen, )["videos"] } if need_video_kwargs: container = av.open(video[0], "r") video_stream = next(stream for stream in container.streams if stream.type == "video") sampling_indices = template_obj.mm_plugin._get_video_sample_indices( video_stream, video_fps, video_maxlen ) total_frames = video_stream.frames video_metadata_kwargs = { "fps": getattr(tokenizer_module["processor"], "video_fps", 24.0), "do_sample_frames": False, "total_num_frames": total_frames, } video_metadata = dict( fps=video_fps, frames_indices=sampling_indices, total_num_frames=total_frames, video_backend="opencv", ) multi_modal_data["video"] = (multi_modal_data["video"], video_metadata) elif batch["audios"][j] is not None: audio = batch["audios"][j] audio_data = template_obj.mm_plugin._regularize_audios( audio, sampling_rate=16000, ) multi_modal_data = {"audio": zip(audio_data["audios"], audio_data["sampling_rates"])} else: multi_modal_data = None vllm_input_data = {"prompt_token_ids": batch["input_ids"][j], "multi_modal_data": multi_modal_data} if "video_metadata_kwargs" in locals() and video_metadata_kwargs is not None: vllm_input_data["mm_processor_kwargs"] = video_metadata_kwargs vllm_inputs.append(vllm_input_data) prompts.append(tokenizer.decode(batch["input_ids"][j], skip_special_tokens=skip_special_tokens)) labels.append( tokenizer.decode( list(filter(lambda x: x != IGNORE_INDEX, batch["labels"][j])), skip_special_tokens=skip_special_tokens, ) ) results = llm.generate(vllm_inputs, sampling_params, lora_request=lora_request) preds = [result.outputs[0].text for result in results] # Accumulate results all_prompts.extend(prompts) all_preds.extend(preds) all_labels.extend(labels) gc.collect() model_predict_end_time = time.time() # Write all results at once outside the loop with open(save_name, "w", encoding="utf-8") as f: for text, pred, label in zip(all_prompts, all_preds, all_labels): f.write(json.dumps({"prompt": text, "predict": pred, "label": label}, ensure_ascii=False) + "\n") print("*" * 70) print(f"{len(all_prompts)} total generated results have been saved at {save_name}.") print("*" * 70) # Write all matrix results when matrix_save_name is not None, # The result matrix is referencing src.llamafactory.train.sft.workflow.run_sft # 127~132 # trainer.save_metrics("predict", predict_results.metrics) # # { # "predict_bleu-4": 4.349975, # "predict_model_preparation_time": 0.0128, # "predict_rouge-1": 21.873359375, # "predict_rouge-2": 4.144340625, # "predict_rouge-l": 10.83949375, # "predict_runtime": 131.664, # "predict_samples_per_second": 0.076, # "predict_steps_per_second": 0.008 # } # if matrix_save_name is not None: predict_time = model_predict_end_time - model_predict_start_time preparation_time = model_predict_start_time - model_preparation_start_time start_time = time.time() dataset = load_dataset("json", data_files=save_name, split="train") dataset = dataset.map(compute_metrics, num_proc=8, remove_columns=dataset.column_names) score_dict = dataset.to_dict() average_score = {} for task, scores in sorted(score_dict.items(), key=lambda x: x[0]): score = sum(scores) / len(scores) if scores else 0.0 print(f"predict_{task}: {score:.4f}") average_score["predict_" + task] = score average_score["predict_model_preparation_time"] = preparation_time average_score["predict_runtime"] = predict_time num_steps = len(range(0, len(train_dataset), batch_size)) average_score["predict_samples_per_second"] = len(dataset) / predict_time if predict_time > 0 else 0.0 average_score["predict_steps_per_second"] = num_steps / predict_time if predict_time > 0 else 0.0 with open(matrix_save_name, "w", encoding="utf-8") as f: json.dump(average_score, f, indent=4) print("*" * 70) print(f"\nDone in {time.time() - start_time:.3f}s.\nScore file saved to {matrix_save_name}.") print("*" * 70) if __name__ == "__main__": fire.Fire(vllm_infer)
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/scripts/loftq_init.py
scripts/loftq_init.py
# Copyright 2025 HuggingFace Inc. and the LlamaFactory team. # # This code is based on the HuggingFace's PEFT library. # https://github.com/huggingface/peft/blob/v0.10.0/examples/loftq_finetuning/quantize_save_load.py # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import os from typing import TYPE_CHECKING import fire from peft import LoftQConfig, LoraConfig, TaskType, get_peft_model from transformers import AutoModelForCausalLM, AutoTokenizer if TYPE_CHECKING: from transformers import PreTrainedModel def quantize_loftq( model_name_or_path: str, output_dir: str, loftq_bits: int = 4, loftq_iter: int = 4, lora_alpha: int = None, lora_rank: int = 16, lora_dropout: float = 0, lora_target: tuple = ("q_proj", "v_proj"), save_safetensors: bool = True, ): r"""Initialize LoRA weights with LoRA-fine-tuning-aware Quantization (LoftQ). Usage: python loftq_init.py --model_name_or_path path_to_model --output_dir output_dir """ if isinstance(lora_target, str): lora_target = [name.strip() for name in lora_target.split(",")] tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, trust_remote_code=True) model = AutoModelForCausalLM.from_pretrained(model_name_or_path, trust_remote_code=True, torch_dtype="auto") loftq_config = LoftQConfig(loftq_bits=loftq_bits, loftq_iter=loftq_iter) lora_config = LoraConfig( task_type=TaskType.CAUSAL_LM, inference_mode=True, r=lora_rank, lora_alpha=lora_alpha if lora_alpha is not None else lora_rank * 2, lora_dropout=lora_dropout, target_modules=lora_target, init_lora_weights="loftq", loftq_config=loftq_config, ) # Init LoftQ model print("Initializing LoftQ weights, it may be take several minutes, wait patiently.") peft_model = get_peft_model(model, lora_config) loftq_dir = os.path.join(output_dir, "loftq_init") # Save LoftQ model setattr(peft_model.peft_config["default"], "base_model_name_or_path", os.path.abspath(output_dir)) setattr(peft_model.peft_config["default"], "init_lora_weights", True) # don't apply loftq again peft_model.save_pretrained(loftq_dir, safe_serialization=save_safetensors) print(f"Adapter weights saved in {loftq_dir}") # Save base model base_model: PreTrainedModel = peft_model.unload() base_model.save_pretrained(output_dir, safe_serialization=save_safetensors) tokenizer.save_pretrained(output_dir) print(f"Model weights saved in {output_dir}") print("- Fine-tune this model with:") print(f"model_name_or_path: {output_dir}") print(f"adapter_name_or_path: {loftq_dir}") print("finetuning_type: lora") print(f"quantization_bit: {loftq_bits}") if __name__ == "__main__": fire.Fire(quantize_loftq)
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/scripts/api_example/test_toolcall.py
scripts/api_example/test_toolcall.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import json import os from openai import OpenAI from transformers.utils.versions import require_version require_version("openai>=1.5.0", "To fix: pip install openai>=1.5.0") def calculate_gpa(grades: list[str], hours: list[int]) -> float: grade_to_score = {"A": 4, "B": 3, "C": 2} total_score, total_hour = 0, 0 for grade, hour in zip(grades, hours): total_score += grade_to_score[grade] * hour total_hour += hour return round(total_score / total_hour, 2) def main(): client = OpenAI( api_key="{}".format(os.getenv("API_KEY", "0")), base_url="http://localhost:{}/v1".format(os.getenv("API_PORT", 8000)), ) tools = [ { "type": "function", "function": { "name": "calculate_gpa", "description": "Calculate the Grade Point Average (GPA) based on grades and credit hours", "parameters": { "type": "object", "properties": { "grades": {"type": "array", "items": {"type": "string"}, "description": "The grades"}, "hours": {"type": "array", "items": {"type": "integer"}, "description": "The credit hours"}, }, "required": ["grades", "hours"], }, }, } ] tool_map = {"calculate_gpa": calculate_gpa} messages = [] messages.append({"role": "user", "content": "My grades are A, A, B, and C. The credit hours are 3, 4, 3, and 2."}) result = client.chat.completions.create(messages=messages, model="test", tools=tools) if result.choices[0].message.tool_calls is None: raise ValueError("Cannot retrieve function call from the response.") messages.append(result.choices[0].message) tool_call = result.choices[0].message.tool_calls[0].function print(tool_call) # Function(arguments='{"grades": ["A", "A", "B", "C"], "hours": [3, 4, 3, 2]}', name='calculate_gpa') name, arguments = tool_call.name, json.loads(tool_call.arguments) tool_result = tool_map[name](**arguments) messages.append({"role": "tool", "content": json.dumps({"gpa": tool_result}, ensure_ascii=False)}) result = client.chat.completions.create(messages=messages, model="test", tools=tools) print(result.choices[0].message.content) # Based on the grades and credit hours you provided, your Grade Point Average (GPA) is 3.42. if __name__ == "__main__": main()
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/scripts/api_example/test_image.py
scripts/api_example/test_image.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import os from openai import OpenAI from transformers.utils.versions import require_version require_version("openai>=1.5.0", "To fix: pip install openai>=1.5.0") def main(): client = OpenAI( api_key="{}".format(os.getenv("API_KEY", "0")), base_url="http://localhost:{}/v1".format(os.getenv("API_PORT", 8000)), ) messages = [] messages.append( { "role": "user", "content": [ {"type": "text", "text": "Output the color and number of each box."}, { "type": "image_url", "image_url": {"url": "https://qianwen-res.oss-cn-beijing.aliyuncs.com/Qwen2-VL/boxes.png"}, }, ], } ) result = client.chat.completions.create(messages=messages, model="test") messages.append(result.choices[0].message) print("Round 1:", result.choices[0].message.content) # The image shows a pyramid of colored blocks with numbers on them. Here are the colors and numbers of ... messages.append( { "role": "user", "content": [ {"type": "text", "text": "What kind of flower is this?"}, { "type": "image_url", "image_url": {"url": "https://qianwen-res.oss-cn-beijing.aliyuncs.com/Qwen2-VL/flowers.jpg"}, }, ], } ) result = client.chat.completions.create(messages=messages, model="test") messages.append(result.choices[0].message) print("Round 2:", result.choices[0].message.content) # The image shows a cluster of forget-me-not flowers. Forget-me-nots are small ... if __name__ == "__main__": main()
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/scripts/convert_ckpt/llamafy_qwen.py
scripts/convert_ckpt/llamafy_qwen.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import json import os from collections import OrderedDict from typing import Any import fire import torch from huggingface_hub import split_torch_state_dict_into_shards from safetensors import safe_open from safetensors.torch import save_file from tqdm import tqdm from transformers.modeling_utils import SAFE_WEIGHTS_INDEX_NAME, SAFE_WEIGHTS_NAME, WEIGHTS_INDEX_NAME, WEIGHTS_NAME from transformers.utils import check_min_version try: check_min_version("4.34.0") except Exception: raise ValueError("Please upgrade `transformers` to 4.34.0") CONFIG_NAME = "config.json" def save_weight(input_dir: str, output_dir: str, shard_size: str, save_safetensors: bool) -> str: qwen_state_dict: dict[str, torch.Tensor] = OrderedDict() for filepath in tqdm(os.listdir(input_dir), desc="Load weights"): if os.path.isfile(os.path.join(input_dir, filepath)) and filepath.endswith(".safetensors"): with safe_open(os.path.join(input_dir, filepath), framework="pt", device="cpu") as f: for key in f.keys(): qwen_state_dict[key] = f.get_tensor(key) llama_state_dict: dict[str, torch.Tensor] = OrderedDict() torch_dtype = None for key, value in tqdm(qwen_state_dict.items(), desc="Convert format"): if torch_dtype is None: torch_dtype = value.dtype if "wte" in key: llama_state_dict["model.embed_tokens.weight"] = value elif "ln_f" in key: llama_state_dict["model.norm.weight"] = value else: key = key.replace("transformer.h", "model.layers") if "attn.c_attn" in key: proj_size = value.size(0) // 3 llama_state_dict[key.replace("attn.c_attn", "self_attn.q_proj")] = value[:proj_size, ...] llama_state_dict[key.replace("attn.c_attn", "self_attn.k_proj")] = value[ proj_size : 2 * proj_size, ... ] llama_state_dict[key.replace("attn.c_attn", "self_attn.v_proj")] = value[2 * proj_size :, ...] elif "attn.c_proj" in key: llama_state_dict[key.replace("attn.c_proj", "self_attn.o_proj")] = value llama_state_dict[key.replace("attn.c_proj.weight", "self_attn.o_proj.bias")] = torch.zeros_like( value[:, 0] ).squeeze() elif "ln_1" in key: llama_state_dict[key.replace("ln_1", "input_layernorm")] = value elif "ln_2" in key: llama_state_dict[key.replace("ln_2", "post_attention_layernorm")] = value elif "mlp.w1" in key: llama_state_dict[key.replace("mlp.w1", "mlp.up_proj")] = value elif "mlp.w2" in key: llama_state_dict[key.replace("mlp.w2", "mlp.gate_proj")] = value elif "mlp.c_proj" in key: llama_state_dict[key.replace("mlp.c_proj", "mlp.down_proj")] = value elif "lm_head" in key: llama_state_dict[key] = value else: raise KeyError(f"Unable to process key {key}") weights_name = SAFE_WEIGHTS_NAME if save_safetensors else WEIGHTS_NAME filename_pattern = weights_name.replace(".bin", "{suffix}.bin").replace(".safetensors", "{suffix}.safetensors") state_dict_split = split_torch_state_dict_into_shards( llama_state_dict, filename_pattern=filename_pattern, max_shard_size=shard_size ) for shard_file, tensors in tqdm(state_dict_split.filename_to_tensors.items(), desc="Save weights"): shard = {tensor: llama_state_dict[tensor].contiguous() for tensor in tensors} if save_safetensors: save_file(shard, os.path.join(output_dir, shard_file), metadata={"format": "pt"}) else: torch.save(shard, os.path.join(output_dir, shard_file)) if not state_dict_split.is_sharded: print(f"Model weights saved in {os.path.join(output_dir, weights_name)}.") else: index = { "metadata": state_dict_split.metadata, "weight_map": state_dict_split.tensor_to_filename, } index_name = SAFE_WEIGHTS_INDEX_NAME if save_safetensors else WEIGHTS_INDEX_NAME with open(os.path.join(output_dir, index_name), "w", encoding="utf-8") as f: json.dump(index, f, indent=2, sort_keys=True) print(f"Model weights saved in {output_dir}.") return str(torch_dtype).replace("torch.", "") def save_config(input_dir: str, output_dir: str, torch_dtype: str): with open(os.path.join(input_dir, CONFIG_NAME), encoding="utf-8") as f: qwen_config_dict: dict[str, Any] = json.load(f) llama2_config_dict: dict[str, Any] = OrderedDict() llama2_config_dict["architectures"] = ["LlamaForCausalLM"] llama2_config_dict["hidden_act"] = "silu" llama2_config_dict["hidden_size"] = qwen_config_dict["hidden_size"] llama2_config_dict["initializer_range"] = qwen_config_dict["initializer_range"] llama2_config_dict["intermediate_size"] = qwen_config_dict["intermediate_size"] // 2 llama2_config_dict["max_position_embeddings"] = qwen_config_dict["max_position_embeddings"] llama2_config_dict["model_type"] = "llama" llama2_config_dict["num_attention_heads"] = qwen_config_dict["num_attention_heads"] llama2_config_dict["num_hidden_layers"] = qwen_config_dict["num_hidden_layers"] llama2_config_dict["num_key_value_heads"] = qwen_config_dict["hidden_size"] // qwen_config_dict["kv_channels"] llama2_config_dict["pretraining_tp"] = 1 llama2_config_dict["rms_norm_eps"] = qwen_config_dict["layer_norm_epsilon"] llama2_config_dict["rope_scaling"] = None llama2_config_dict["tie_word_embeddings"] = qwen_config_dict["tie_word_embeddings"] llama2_config_dict["torch_dtype"] = torch_dtype llama2_config_dict["transformers_version"] = "4.34.0" llama2_config_dict["use_cache"] = True llama2_config_dict["vocab_size"] = qwen_config_dict["vocab_size"] llama2_config_dict["attention_bias"] = True with open(os.path.join(output_dir, CONFIG_NAME), "w", encoding="utf-8") as f: json.dump(llama2_config_dict, f, indent=2) print(f"Model config saved in {os.path.join(output_dir, CONFIG_NAME)}") def llamafy_qwen( input_dir: str, output_dir: str, shard_size: str = "2GB", save_safetensors: bool = False, ): r"""Convert the Qwen models in the same format as LLaMA2. Usage: python llamafy_qwen.py --input_dir input --output_dir output Converted model: https://huggingface.co/hiyouga/Qwen-14B-Chat-LLaMAfied """ try: os.makedirs(output_dir, exist_ok=False) except Exception as e: raise print("Output dir already exists", e) torch_dtype = save_weight(input_dir, output_dir, shard_size, save_safetensors) save_config(input_dir, output_dir, torch_dtype) if __name__ == "__main__": fire.Fire(llamafy_qwen)
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/scripts/convert_ckpt/tiny_llama4.py
scripts/convert_ckpt/tiny_llama4.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from transformers import Llama4Config, Llama4ForConditionalGeneration, Llama4TextConfig, Llama4VisionConfig if __name__ == "__main__": vision_config = Llama4VisionConfig( hidden_size=1408, image_size=336, intermediate_size=5632, num_attention_heads=16, num_hidden_layers=4, vision_output_dim=4096, ) text_config = Llama4TextConfig( hidden_size=512, intermediate_size=1024, intermediate_size_mlp=1024, num_hidden_layers=4, num_attention_heads=8, num_key_value_heads=2, head_dim=512 // 8, num_local_experts=2, ) config = Llama4Config(vision_config=vision_config, text_config=text_config) model = Llama4ForConditionalGeneration._from_config(config) model.save_pretrained("tiny-llama4")
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/scripts/convert_ckpt/llamafy_baichuan2.py
scripts/convert_ckpt/llamafy_baichuan2.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import json import os from collections import OrderedDict from typing import Any import fire import torch from huggingface_hub import split_torch_state_dict_into_shards from safetensors.torch import save_file from tqdm import tqdm from transformers.modeling_utils import SAFE_WEIGHTS_INDEX_NAME, SAFE_WEIGHTS_NAME, WEIGHTS_INDEX_NAME, WEIGHTS_NAME CONFIG_NAME = "config.json" def save_weight(input_dir: str, output_dir: str, shard_size: str, save_safetensors: bool): baichuan2_state_dict: dict[str, torch.Tensor] = OrderedDict() for filepath in tqdm(os.listdir(input_dir), desc="Load weights"): if os.path.isfile(os.path.join(input_dir, filepath)) and filepath.endswith(".bin"): shard_weight = torch.load(os.path.join(input_dir, filepath), map_location="cpu", weights_only=True) baichuan2_state_dict.update(shard_weight) llama_state_dict: dict[str, torch.Tensor] = OrderedDict() for key, value in tqdm(baichuan2_state_dict.items(), desc="Convert format"): if "W_pack" in key: proj_size = value.size(0) // 3 llama_state_dict[key.replace("W_pack", "q_proj")] = value[:proj_size, :] llama_state_dict[key.replace("W_pack", "k_proj")] = value[proj_size : 2 * proj_size, :] llama_state_dict[key.replace("W_pack", "v_proj")] = value[2 * proj_size :, :] elif "lm_head" in key: llama_state_dict[key] = torch.nn.functional.normalize(value) else: llama_state_dict[key] = value weights_name = SAFE_WEIGHTS_NAME if save_safetensors else WEIGHTS_NAME filename_pattern = weights_name.replace(".bin", "{suffix}.bin").replace(".safetensors", "{suffix}.safetensors") state_dict_split = split_torch_state_dict_into_shards( llama_state_dict, filename_pattern=filename_pattern, max_shard_size=shard_size ) for shard_file, tensors in tqdm(state_dict_split.filename_to_tensors.items(), desc="Save weights"): shard = {tensor: llama_state_dict[tensor].contiguous() for tensor in tensors} if save_safetensors: save_file(shard, os.path.join(output_dir, shard_file), metadata={"format": "pt"}) else: torch.save(shard, os.path.join(output_dir, shard_file)) if not state_dict_split.is_sharded: print(f"Model weights saved in {os.path.join(output_dir, weights_name)}.") else: index = { "metadata": state_dict_split.metadata, "weight_map": state_dict_split.tensor_to_filename, } index_name = SAFE_WEIGHTS_INDEX_NAME if save_safetensors else WEIGHTS_INDEX_NAME with open(os.path.join(output_dir, index_name), "w", encoding="utf-8") as f: json.dump(index, f, indent=2, sort_keys=True) print(f"Model weights saved in {output_dir}.") def save_config(input_dir: str, output_dir: str): with open(os.path.join(input_dir, CONFIG_NAME), encoding="utf-8") as f: llama2_config_dict: dict[str, Any] = json.load(f) llama2_config_dict["architectures"] = ["LlamaForCausalLM"] llama2_config_dict.pop("auto_map", None) llama2_config_dict.pop("tokenizer_class", None) llama2_config_dict["model_type"] = "llama" with open(os.path.join(output_dir, CONFIG_NAME), "w", encoding="utf-8") as f: json.dump(llama2_config_dict, f, indent=2) print(f"Model config saved in {os.path.join(output_dir, CONFIG_NAME)}") def llamafy_baichuan2( input_dir: str, output_dir: str, shard_size: str = "2GB", save_safetensors: bool = True, ): r"""Convert the Baichuan2-7B model in the same format as LLaMA2-7B. Usage: python llamafy_baichuan2.py --input_dir input --output_dir output Converted model: https://huggingface.co/hiyouga/Baichuan2-7B-Base-LLaMAfied """ try: os.makedirs(output_dir, exist_ok=False) except Exception as e: raise print("Output dir already exists", e) save_weight(input_dir, output_dir, shard_size, save_safetensors) save_config(input_dir, output_dir) if __name__ == "__main__": fire.Fire(llamafy_baichuan2)
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/scripts/stat_utils/cal_mfu.py
scripts/stat_utils/cal_mfu.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import json import os import fire import torch import torch.distributed as dist from transformers import AutoConfig from llamafactory.train.tuner import run_exp BASE = 2 # gemm (add + mul) def compute_model_flops( model_name_or_path: str, total_batch_size: int, seq_length: int, include_backward: bool = True, include_recompute: bool = False, include_flashattn: bool = False, ) -> int: r"""Calculate the FLOPs of model per forward/backward pass.""" config = AutoConfig.from_pretrained(model_name_or_path) hidden_size = getattr(config, "hidden_size", None) vocab_size = getattr(config, "vocab_size", None) intermediate_size = getattr(config, "intermediate_size", None) num_attention_heads = getattr(config, "num_attention_heads", None) num_key_value_heads = getattr(config, "num_key_value_heads", None) num_hidden_layers = getattr(config, "num_hidden_layers", None) tie_word_embeddings = getattr(config, "tie_word_embeddings", False) # mlp module mlp_flops_per_token = 3 * BASE * hidden_size * intermediate_size # up, gate, down mlp_flops = total_batch_size * seq_length * num_hidden_layers * mlp_flops_per_token # attn projector module q_flops_per_token = BASE * hidden_size * hidden_size o_flops_per_token = BASE * hidden_size * hidden_size k_flops_per_token = BASE * hidden_size * hidden_size * num_key_value_heads // num_attention_heads v_flops_per_token = BASE * hidden_size * hidden_size * num_key_value_heads // num_attention_heads attn_proj_flops_per_token = q_flops_per_token + o_flops_per_token + k_flops_per_token + v_flops_per_token attn_proj_flops = total_batch_size * seq_length * num_hidden_layers * attn_proj_flops_per_token # attn sdpa module sdpa_flops_per_layer = 2 * BASE * hidden_size * seq_length * seq_length # (q * k^T) * v sdpa_flops = total_batch_size * num_hidden_layers * sdpa_flops_per_layer # embedding module embedding_flops_per_token = hidden_size * vocab_size embedding_flops = total_batch_size * seq_length * embedding_flops_per_token if tie_word_embeddings is False: embedding_flops *= 2 non_embedding_flops = mlp_flops + attn_proj_flops + sdpa_flops non_embedding_coeff, embedding_coeff = 1, 1 if include_backward: non_embedding_coeff += 2 embedding_coeff += 2 if include_recompute: non_embedding_coeff += 1 total_flops = non_embedding_coeff * non_embedding_flops + embedding_coeff * embedding_flops if include_flashattn: total_flops += sdpa_flops return total_flops def compute_device_flops(world_size: int) -> float: r"""Calculate the FLOPs of the device capability per second.""" device_name = torch.cuda.get_device_name() if "H100" in device_name or "H800" in device_name: return 989 * 1e12 * world_size elif "A100" in device_name or "A800" in device_name: return 312 * 1e12 * world_size elif "V100" in device_name: return 125 * 1e12 * world_size elif "4090" in device_name: return 98 * 1e12 * world_size else: raise NotImplementedError(f"Device not supported: {device_name}.") def calculate_mfu( model_name_or_path: str, batch_size: int = 1, seq_length: int = 1024, num_steps: int = 100, finetuning_type: str = "lora", flash_attn: str = "auto", deepspeed_stage: int = 0, disable_gc: bool = False, liger_kernel: bool = False, unsloth_gc: bool = False, ) -> float: r"""Calculate MFU for given model and hyper-params. Usage: python cal_mfu.py --model_name_or_path path_to_model --batch_size 1 --seq_length 1024 """ args = { "model_name_or_path": model_name_or_path, "flash_attn": flash_attn, "disable_gradient_checkpointing": disable_gc, "enable_liger_kernel": liger_kernel, "use_unsloth_gc": unsloth_gc, "stage": "pt", "do_train": True, "finetuning_type": finetuning_type, "dataset": "c4_demo", "cutoff_len": seq_length, "output_dir": os.path.join("saves", "test_mfu"), "logging_strategy": "no", "save_strategy": "no", "save_only_model": True, "overwrite_output_dir": True, "per_device_train_batch_size": batch_size, "max_steps": num_steps, "bf16": True, } if deepspeed_stage in [2, 3]: args["deepspeed"] = f"examples/deepspeed/ds_z{deepspeed_stage}_config.json" run_exp(args) if dist.is_initialized(): dist.barrier() world_size = dist.get_world_size() else: world_size = 1 if int(os.getenv("LOCAL_RANK", "0")) == 0: with open(os.path.join("saves", "test_mfu", "all_results.json"), encoding="utf-8") as f: result = json.load(f) total_batch_size = batch_size * world_size mfu_value = ( result["train_steps_per_second"] * compute_model_flops(model_name_or_path, total_batch_size, seq_length) / compute_device_flops(world_size) ) print(f"MFU: {mfu_value * 100:.2f}%") if __name__ == "__main__": fire.Fire(calculate_mfu)
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/scripts/stat_utils/length_cdf.py
scripts/stat_utils/length_cdf.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from collections import defaultdict import fire from tqdm import tqdm from llamafactory.data import get_dataset, get_template_and_fix_tokenizer from llamafactory.hparams import get_train_args from llamafactory.model import load_tokenizer def length_cdf( model_name_or_path: str, dataset: str = "alpaca_en_demo", dataset_dir: str = "data", template: str = "default", interval: int = 1000, ): r"""Calculate the distribution of the input lengths in the dataset. Usage: export CUDA_VISIBLE_DEVICES=0 python length_cdf.py --model_name_or_path path_to_model --dataset alpaca_en_demo --template default """ model_args, data_args, training_args, _, _ = get_train_args( dict( stage="sft", model_name_or_path=model_name_or_path, dataset=dataset, dataset_dir=dataset_dir, template=template, cutoff_len=1_000_000, preprocessing_num_workers=16, output_dir="dummy_dir", overwrite_cache=True, do_train=True, ) ) tokenizer_module = load_tokenizer(model_args) template = get_template_and_fix_tokenizer(tokenizer_module["tokenizer"], data_args) trainset = get_dataset(template, model_args, data_args, training_args, "sft", **tokenizer_module)["train_dataset"] total_num = len(trainset) length_dict = defaultdict(int) for sample in tqdm(trainset["input_ids"], desc="Collecting lengths"): length_dict[len(sample) // interval * interval] += 1 length_tuples = list(length_dict.items()) length_tuples.sort() count_accu, prob_accu = 0, 0 for length, count in length_tuples: count_accu += count prob_accu += count / total_num * 100 print(f"{count_accu:d} ({prob_accu:.2f}%) samples have length < {length + interval}.") if __name__ == "__main__": fire.Fire(length_cdf)
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/scripts/stat_utils/cal_lr.py
scripts/stat_utils/cal_lr.py
# Copyright 2025 imoneoi and the LlamaFactory team. # # This code is inspired by the imoneoi's OpenChat library. # https://github.com/imoneoi/openchat/blob/3.6.0/ochat/training_deepspeed/train.py # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import math from typing import Literal import fire import torch from torch.utils.data import DataLoader from tqdm import tqdm from transformers import DataCollatorForLanguageModeling from llamafactory.data import MultiModalDataCollatorForSeq2Seq, get_dataset, get_template_and_fix_tokenizer from llamafactory.extras.constants import IGNORE_INDEX from llamafactory.hparams import get_train_args from llamafactory.model import load_tokenizer BASE_LR = 3e-4 # 1.5e-4 for 30B-70B models BASE_BS = 4_000_000 # from llama paper def calculate_lr( model_name_or_path: str, batch_size: int, # total batch size, namely (batch size * gradient accumulation * world size) stage: Literal["pt", "sft"] = "sft", dataset: str = "alpaca_en_demo", dataset_dir: str = "data", template: str = "default", cutoff_len: int = 2048, # i.e. maximum input length during training is_mistral_or_gemma: bool = False, # mistral and gemma models opt for a smaller learning rate, packing: bool = False, ): r"""Calculate the optimal learning rate for 7B/13B models using LLaMA's hyper-parameters. Usage: python cal_lr.py --model_name_or_path path_to_model --dataset alpaca_en_demo --cutoff_len 1024 --batch_size 16 """ model_args, data_args, training_args, _, _ = get_train_args( dict( stage=stage, model_name_or_path=model_name_or_path, dataset=dataset, dataset_dir=dataset_dir, template=template, cutoff_len=cutoff_len, packing=packing, preprocessing_num_workers=16, output_dir="dummy_dir", overwrite_cache=True, do_train=True, ) ) tokenizer_module = load_tokenizer(model_args) tokenizer = tokenizer_module["tokenizer"] template = get_template_and_fix_tokenizer(tokenizer, data_args) trainset = get_dataset(template, model_args, data_args, training_args, stage, **tokenizer_module)["train_dataset"] if stage == "pt": data_collator = DataCollatorForLanguageModeling(tokenizer=tokenizer, mlm=False) elif stage == "sft": data_collator = MultiModalDataCollatorForSeq2Seq( template=template, tokenizer=tokenizer, label_pad_token_id=IGNORE_INDEX ) else: raise NotImplementedError(f"Stage does not supported: {stage}.") dataloader = DataLoader(trainset, batch_size, shuffle=False, collate_fn=data_collator, pin_memory=True) valid_tokens, total_tokens = 0, 0 for batch in tqdm(dataloader, desc="Collecting valid tokens"): valid_tokens += torch.sum(batch["labels"] != IGNORE_INDEX).item() total_tokens += torch.numel(batch["labels"]) valid_ratio = valid_tokens / total_tokens token_batch_size = cutoff_len * batch_size * valid_ratio lr = BASE_LR * math.sqrt(token_batch_size / BASE_BS) # lr ~ sqrt(batch_size) lr = lr / 6.0 if is_mistral_or_gemma else lr print( f"Optimal learning rate is {lr:.2e} for valid ratio% {valid_ratio * 100:.2f} " f"and effective token batch size {token_batch_size:.2f}" ) if __name__ == "__main__": fire.Fire(calculate_lr)
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/scripts/stat_utils/cal_flops.py
scripts/stat_utils/cal_flops.py
# Copyright 2025 Microsoft Corporation and the LlamaFactory team. # # This code is inspired by the Microsoft's DeepSpeed library. # https://www.deepspeed.ai/tutorials/flops-profiler/ # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import fire import torch from deepspeed.accelerator import get_accelerator # type: ignore from deepspeed.profiling.flops_profiler import get_model_profile # type: ignore from llamafactory.chat import ChatModel def calculate_flops( model_name_or_path: str, batch_size: int = 1, seq_length: int = 512, flash_attn: str = "auto", ): r"""Calculate the flops of pre-trained models. Usage: python cal_flops.py --model_name_or_path path_to_model --batch_size 1 --seq_length 512 """ with get_accelerator().device(0): chat_model = ChatModel(dict(model_name_or_path=model_name_or_path, template="empty", flash_attn=flash_attn)) fake_input = torch.ones((batch_size, seq_length), dtype=torch.long, device=chat_model.engine.model.device) input_dict = {"input_ids": fake_input, "labels": fake_input.clone()} flops, macs, params = get_model_profile( chat_model.engine.model, kwargs=input_dict, print_profile=True, detailed=True ) print("FLOPs:", flops) print("MACs:", macs) print("Params:", params) if __name__ == "__main__": fire.Fire(calculate_flops)
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/scripts/stat_utils/cal_ppl.py
scripts/stat_utils/cal_ppl.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import json from dataclasses import dataclass from typing import Any, Literal import fire import torch from torch.utils.data import DataLoader from tqdm import tqdm from transformers import DataCollatorForLanguageModeling from llamafactory.data import MultiModalDataCollatorForSeq2Seq, get_dataset, get_template_and_fix_tokenizer from llamafactory.extras.constants import IGNORE_INDEX from llamafactory.hparams import get_train_args from llamafactory.model import load_model, load_tokenizer @dataclass class PairwiseDataCollatorWithPadding(MultiModalDataCollatorForSeq2Seq): r"""Data collator for pairwise data.""" train_on_prompt: bool = False def __call__(self, features: list[dict[str, Any]]) -> dict[str, torch.Tensor]: r"""Pad batched data to the longest sequence in the batch.""" chosen_features = [] for feature in features: chosen_features.append( { "input_ids": feature["chosen_input_ids"], "attention_mask": feature["chosen_attention_mask"], "labels": feature["chosen_input_ids"] if self.train_on_prompt else feature["chosen_labels"], "images": feature["images"], "videos": feature["videos"], "audios": feature["audios"], } ) return super().__call__(chosen_features) def calculate_ppl( model_name_or_path: str, save_name: str = "ppl.json", batch_size: int = 4, stage: Literal["pt", "sft", "rm"] = "sft", dataset: str = "alpaca_en_demo", dataset_dir: str = "data", template: str = "default", cutoff_len: int = 2048, max_samples: int | None = None, train_on_prompt: bool = False, ): r"""Calculate the ppl on the dataset of the pre-trained models. Usage: export CUDA_VISIBLE_DEVICES=0 python cal_ppl.py --model_name_or_path path_to_model --dataset alpaca_en_demo --save_name ppl.json """ model_args, data_args, training_args, finetuning_args, _ = get_train_args( dict( stage=stage, model_name_or_path=model_name_or_path, dataset=dataset, dataset_dir=dataset_dir, template=template, cutoff_len=cutoff_len, max_samples=max_samples, train_on_prompt=train_on_prompt, preprocessing_num_workers=16, output_dir="dummy_dir", overwrite_cache=True, do_train=True, ) ) tokenizer_module = load_tokenizer(model_args) tokenizer = tokenizer_module["tokenizer"] template = get_template_and_fix_tokenizer(tokenizer, data_args) trainset = get_dataset(template, model_args, data_args, training_args, stage, **tokenizer_module)["train_dataset"] model = load_model(tokenizer, model_args, finetuning_args, is_trainable=False) if stage == "pt": data_collator = DataCollatorForLanguageModeling(tokenizer=tokenizer, mlm=False) elif stage == "sft": data_collator = MultiModalDataCollatorForSeq2Seq( template=template, tokenizer=tokenizer, label_pad_token_id=IGNORE_INDEX ) elif stage == "rm": data_collator = PairwiseDataCollatorWithPadding( template=template, tokenizer=tokenizer, label_pad_token_id=IGNORE_INDEX, train_on_prompt=train_on_prompt ) else: raise NotImplementedError(f"Stage does not supported: {stage}.") dataloader = DataLoader(trainset, batch_size, shuffle=False, collate_fn=data_collator, pin_memory=True) criterion = torch.nn.CrossEntropyLoss(reduction="none") total_ppl = 0 perplexities = [] batch: dict[str, torch.Tensor] with torch.no_grad(): for batch in tqdm(dataloader, desc="Computing perplexities"): batch = batch.to(model.device) outputs = model(**batch) shift_logits: torch.Tensor = outputs["logits"][..., :-1, :] shift_labels: torch.Tensor = batch["labels"][..., 1:] loss_mask = shift_labels != IGNORE_INDEX flatten_logits = shift_logits.contiguous().view(shift_labels.size(0) * shift_labels.size(1), -1) flatten_labels = shift_labels.contiguous().view(-1) token_logps: torch.Tensor = criterion(flatten_logits, flatten_labels) token_logps = token_logps.contiguous().view(shift_logits.size(0), -1) sentence_logps = (token_logps * loss_mask).sum(-1) / loss_mask.sum(-1) total_ppl += sentence_logps.exp().sum().item() perplexities.extend(sentence_logps.exp().tolist()) with open(save_name, "w", encoding="utf-8") as f: json.dump(perplexities, f, indent=2) print(f"Average perplexity is {total_ppl / len(perplexities):.2f}") print(f"Perplexities have been saved at {save_name}.") if __name__ == "__main__": fire.Fire(calculate_ppl)
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/tests_v1/conftest.py
tests_v1/conftest.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """LlamaFactory test configuration. Contains shared fixtures, pytest configuration, and custom markers. """ import os import sys import pytest import torch import torch.distributed as dist from pytest import Config, FixtureRequest, Item, MonkeyPatch from llamafactory.v1.accelerator.helper import get_current_accelerator, get_device_count from llamafactory.v1.utils.env import is_env_enabled from llamafactory.v1.utils.packages import is_transformers_version_greater_than CURRENT_DEVICE = get_current_accelerator().type def pytest_configure(config: Config): """Register custom pytest markers.""" config.addinivalue_line( "markers", "slow: marks tests as slow (deselect with '-m \"not slow\"' or set RUN_SLOW=1 to run)", ) config.addinivalue_line( "markers", "runs_on: test requires specific device type, e.g., @pytest.mark.runs_on(['cuda'])", ) config.addinivalue_line( "markers", "require_distributed(num_devices): allow multi-device execution (default: 2)", ) def _handle_runs_on(items: list[Item]): """Skip tests on specified device TYPES (cpu/cuda/npu).""" for item in items: marker = item.get_closest_marker("runs_on") if not marker: continue devices = marker.args[0] if isinstance(devices, str): devices = [devices] if CURRENT_DEVICE not in devices: item.add_marker(pytest.mark.skip(reason=f"test requires one of {devices} (current: {CURRENT_DEVICE})")) def _handle_slow_tests(items: list[Item]): """Skip slow tests unless RUN_SLOW is enabled.""" if not is_env_enabled("RUN_SLOW"): skip_slow = pytest.mark.skip(reason="slow test (set RUN_SLOW=1 to run)") for item in items: if "slow" in item.keywords: item.add_marker(skip_slow) def _get_visible_devices_env() -> str | None: """Return device visibility env var name.""" if CURRENT_DEVICE == "cuda": return "CUDA_VISIBLE_DEVICES" elif CURRENT_DEVICE == "npu": return "ASCEND_RT_VISIBLE_DEVICES" else: return None def _handle_device_visibility(items: list[Item]): """Handle device visibility based on test markers.""" env_key = _get_visible_devices_env() if env_key is None or CURRENT_DEVICE in ("cpu", "mps"): return # Parse visible devices visible_devices_env = os.environ.get(env_key) if visible_devices_env is None: available = get_device_count() else: visible_devices = [v for v in visible_devices_env.split(",") if v != ""] available = len(visible_devices) for item in items: marker = item.get_closest_marker("require_distributed") if not marker: continue required = marker.args[0] if marker.args else 2 if available < required: item.add_marker(pytest.mark.skip(reason=f"test requires {required} devices, but only {available} visible")) def pytest_collection_modifyitems(config: Config, items: list[Item]): """Modify test collection based on markers and environment.""" # Handle version compatibility (from HEAD) skip_bc = pytest.mark.skip(reason="Skip backward compatibility tests") for item in items: if "tests_v1" in str(item.fspath) and not is_transformers_version_greater_than("4.57.0"): item.add_marker(skip_bc) _handle_slow_tests(items) _handle_runs_on(items) _handle_device_visibility(items) @pytest.fixture(autouse=True) def _cleanup_distributed_state(): """Cleanup distributed state after each test.""" yield if dist.is_initialized(): dist.destroy_process_group() @pytest.fixture(autouse=True) def _manage_distributed_env(request: FixtureRequest, monkeypatch: MonkeyPatch) -> None: """Set environment variables for distributed tests if specific devices are requested.""" env_key = _get_visible_devices_env() if not env_key: return # Save old environment for logic checks, monkeypatch handles restoration old_value = os.environ.get(env_key) marker = request.node.get_closest_marker("require_distributed") if marker: # distributed test required = marker.args[0] if marker.args else 2 specific_devices = marker.args[1] if len(marker.args) > 1 else None if specific_devices: devices_str = ",".join(map(str, specific_devices)) else: devices_str = ",".join(str(i) for i in range(required)) monkeypatch.setenv(env_key, devices_str) # add project root dir to path for mp run project_root = os.path.abspath(os.path.join(os.path.dirname(__file__), "..")) if project_root not in sys.path: sys.path.insert(0, project_root) os.environ["PYTHONPATH"] = project_root + os.pathsep + os.environ.get("PYTHONPATH", "") else: # non-distributed test if old_value: visible_devices = [v for v in old_value.split(",") if v != ""] monkeypatch.setenv(env_key, visible_devices[0] if visible_devices else "0") else: monkeypatch.setenv(env_key, "0") if CURRENT_DEVICE == "cuda": monkeypatch.setattr(torch.cuda, "device_count", lambda: 1) elif CURRENT_DEVICE == "npu": monkeypatch.setattr(torch.npu, "device_count", lambda: 1)
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/tests_v1/utils/test_batching_queue.py
tests_v1/utils/test_batching_queue.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from llamafactory.v1.utils.batching_queue import DynamicBatchSizeBuffer, TextBatchingQueue def create_sample(length: int): """Helper to create a mock sample with a specific token length.""" return {"input_ids": torch.ones(length), "attention_mask": torch.ones(length)} class TestDynamicBatchSizeBuffer: def test_append_and_token_count(self): buffer = DynamicBatchSizeBuffer() buffer.append(create_sample(10)) buffer.append(create_sample(20)) assert len(buffer) == 2 assert buffer.total_token_count == 30 def test_get_samples_within_budget(self): buffer = DynamicBatchSizeBuffer() buffer.append(create_sample(10)) buffer.append(create_sample(10)) buffer.append(create_sample(50)) # This one is large # Request 25 tokens. Should get the first two (20 tokens total) samples = buffer.get_samples(max_tokens_per_iteration=25) assert len(samples) == 2 def test_force_return_first_sample(self): buffer = DynamicBatchSizeBuffer() buffer.append(create_sample(100)) # Even though budget is 50, force=True (default) should return the 100-token sample samples = buffer.get_samples(max_tokens_per_iteration=50, force=True) assert len(samples) == 1 assert len(samples[0]["input_ids"]) == 100 def test_flush_removes_used_samples(self): buffer = DynamicBatchSizeBuffer() buffer.append(create_sample(10)) buffer.append(create_sample(20)) # Take the first sample buffer.get_samples(max_tokens_per_iteration=15) buffer.flush() assert len(buffer) == 1 assert buffer.total_token_count == 20 # The remaining sample should now be at the start remaining = buffer.get_samples(max_tokens_per_iteration=50) assert len(remaining[0]["input_ids"]) == 20 class TestTextBatchingQueue: def test_is_full_filled(self): queue = TextBatchingQueue(token_micro_bsz=100, buffer_size=2) queue.put_item(create_sample(10)) assert not queue.is_full_filled() # Only 1 sample, buffer_size=2 queue.put_item(create_sample(10)) assert not queue.is_full_filled() # 2 samples, but only 20 tokens (min 100) queue.put_item(create_sample(90)) assert queue.is_full_filled() # Meets both conditions def test_warmup_logic(self): # token_micro_bsz=1000, starts at 200, reaches 1000 at step 10 queue = TextBatchingQueue(token_micro_bsz=1000, bsz_warmup_steps=10, bsz_warmup_init_mbtoken=200) # Step 0: should be init value assert queue.get_cur_token_micro_bsz() == 200 # Step 5: halfway through warmup (200 + (800 * 5/10)) = 600 queue._step = 5 assert queue.get_cur_token_micro_bsz() == 600 # Step 11: past warmup queue._step = 11 assert queue.get_cur_token_micro_bsz() == 1000 def test_get_micro_batch_integration(self): queue = TextBatchingQueue(token_micro_bsz=50, buffer_size=1) queue.put_item(create_sample(20)) queue.put_item(create_sample(20)) queue.put_item(create_sample(20)) # At step 0 (warmup not triggered as bsz_warmup_steps is -1 default), # it should take samples up to 50 tokens. batch = queue.get_micro_batch(step=0) assert len(batch) == 2 assert queue.empty() is False batch_2 = queue.get_micro_batch(step=1) assert len(batch_2) == 1 assert queue.empty() is True
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/tests_v1/plugins/data_plugins/test_converter.py
tests_v1/plugins/data_plugins/test_converter.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import random import pytest from datasets import load_dataset from llamafactory.v1.config.data_args import DataArguments from llamafactory.v1.core.data_engine import DataEngine from llamafactory.v1.plugins.data_plugins.converter import DataConverterPlugin @pytest.mark.parametrize("num_samples", [16]) def test_alpaca_converter(num_samples: int): data_args = DataArguments(dataset="llamafactory/v1-dataset-info/tiny-supervised-dataset.yaml") data_engine = DataEngine(data_args) original_data = load_dataset("llamafactory/tiny-supervised-dataset", split="train") indexes = random.choices(range(len(data_engine)), k=num_samples) for index in indexes: print(data_engine[index]) expected_data = { "messages": [ { "role": "user", "content": [ {"type": "text", "value": original_data[index]["instruction"] + original_data[index]["input"]} ], "loss_weight": 0.0, }, { "role": "assistant", "content": [{"type": "text", "value": original_data[index]["output"]}], "loss_weight": 1.0, }, ] } assert data_engine[index] == {"_dataset_name": "tiny_dataset", **expected_data} def test_sharegpt_converter(): example = { "conversations": [ {"from": "system", "value": "System"}, {"from": "human", "value": "User"}, {"from": "function_call", "value": "Tool"}, {"from": "observation", "value": "Observation"}, {"from": "gpt", "value": "Assistant"}, ] } expected_data = { "messages": [ {"content": [{"type": "text", "value": "System"}], "loss_weight": 0.0, "role": "system"}, {"content": [{"type": "text", "value": "User"}], "loss_weight": 0.0, "role": "user"}, {"content": [{"type": "tool_calls", "value": "Tool"}], "loss_weight": 1.0, "role": "assistant"}, {"content": [{"type": "text", "value": "Observation"}], "loss_weight": 0.0, "role": "tool"}, {"content": [{"type": "text", "value": "Assistant"}], "loss_weight": 1.0, "role": "assistant"}, ] } assert DataConverterPlugin("sharegpt")(example) == expected_data @pytest.mark.parametrize("num_samples", [16]) def test_pair_converter(num_samples: int): data_args = DataArguments(dataset="llamafactory/v1-dataset-info/orca-dpo-pairs.yaml") data_engine = DataEngine(data_args) original_data = load_dataset("HuggingFaceH4/orca_dpo_pairs", split="train_prefs") indexes = random.choices(range(len(data_engine)), k=num_samples) for index in indexes: print(data_engine[index]) print(original_data[index]) expected_data = { "chosen_messages": [ { "role": "system", "content": [{"type": "text", "value": original_data[index]["chosen"][0]["content"]}], "loss_weight": 0.0, }, { "role": "user", "content": [{"type": "text", "value": original_data[index]["chosen"][1]["content"]}], "loss_weight": 0.0, }, { "role": "assistant", "content": [{"type": "text", "value": original_data[index]["chosen"][2]["content"]}], "loss_weight": 1.0, }, ], "rejected_messages": [ { "role": "system", "content": [{"type": "text", "value": original_data[index]["rejected"][0]["content"]}], "loss_weight": 0.0, }, { "role": "user", "content": [{"type": "text", "value": original_data[index]["rejected"][1]["content"]}], "loss_weight": 0.0, }, { "role": "assistant", "content": [{"type": "text", "value": original_data[index]["rejected"][2]["content"]}], "loss_weight": 1.0, }, ], } assert data_engine[index] == {"_dataset_name": "tiny_dataset", **expected_data} if __name__ == "__main__": test_alpaca_converter(1) test_sharegpt_converter() test_pair_converter(1)
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/tests_v1/plugins/model_plugins/test_kernel_plugin.py
tests_v1/plugins/model_plugins/test_kernel_plugin.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import sys from unittest.mock import MagicMock, patch import pytest from transformers import AutoModelForCausalLM from llamafactory.v1.accelerator.helper import get_current_accelerator @pytest.fixture(autouse=True) def clear_accelerator_cache(): get_current_accelerator.cache_clear() def reload_kernels(): """Helper to reload kernel modules to respect mocked accelerator.""" # Unload kernel interface and registry keys_to_remove = [k for k in sys.modules if k.startswith("llamafactory.v1.plugins.model_plugins.kernels")] for k in keys_to_remove: del sys.modules[k] @patch("torch.accelerator.current_accelerator") def test_apply_kernel(mock_get_accelerator: MagicMock): mock_device = MagicMock() setattr(mock_device, "type", "npu") mock_get_accelerator.return_value = mock_device # Force reload of kernels with mocked accelerator reload_kernels() from llamafactory.v1.plugins.model_plugins.kernels.interface import apply_default_kernels model = AutoModelForCausalLM.from_pretrained("llamafactory/tiny-random-qwen2.5") original_rmsnorm_forward = model.model.layers[0].input_layernorm.forward original_swiglu_forward = model.model.layers[0].mlp.forward model = apply_default_kernels(model=model, include_kernels="npu_fused_rmsnorm") assert model.model.layers[0].input_layernorm.forward.__func__ is not original_rmsnorm_forward.__func__ assert model.model.layers[0].mlp.forward.__func__ is original_swiglu_forward.__func__ @patch("torch.accelerator.current_accelerator") def test_apply_all_kernels(mock_get_accelerator: MagicMock): get_current_accelerator.cache_clear() mock_device = MagicMock() setattr(mock_device, "type", "npu") mock_get_accelerator.return_value = mock_device # Force reload of kernels with mocked accelerator reload_kernels() from llamafactory.v1.plugins.model_plugins.kernels.interface import apply_default_kernels model = AutoModelForCausalLM.from_pretrained("llamafactory/tiny-random-qwen2.5") original_rmsnorm_forward = model.model.layers[0].input_layernorm.forward original_swiglu_forward = model.model.layers[0].mlp.forward model = apply_default_kernels(model=model, include_kernels=True) assert model.model.layers[0].input_layernorm.forward.__func__ is not original_rmsnorm_forward.__func__ assert model.model.layers[0].mlp.forward.__func__ is not original_swiglu_forward.__func__
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/tests_v1/plugins/model_plugins/test_init_plugin.py
tests_v1/plugins/model_plugins/test_init_plugin.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import pytest from llamafactory.v1.accelerator.interface import DistributedInterface from llamafactory.v1.config.arg_parser import get_args from llamafactory.v1.core.model_loader import ModelLoader def test_init_on_meta(): _, model_args, *_ = get_args( dict( model="llamafactory/tiny-random-qwen2.5", init_config={"name": "init_on_meta"}, ) ) model_loader = ModelLoader(model_args=model_args) assert model_loader.model.device.type == "meta" @pytest.mark.runs_on(["cuda", "npu"]) def test_init_on_rank0(): _, model_args, *_ = get_args( dict( model="llamafactory/tiny-random-qwen2.5", init_config={"name": "init_on_rank0"}, ) ) model_loader = ModelLoader(model_args=model_args) if DistributedInterface().get_rank() == 0: assert model_loader.model.device.type == "cpu" else: assert model_loader.model.device.type == "meta" def test_init_on_default(): _, model_args, *_ = get_args( dict( model="llamafactory/tiny-random-qwen2.5", init_config={"name": "init_on_default"}, ) ) model_loader = ModelLoader(model_args=model_args) assert model_loader.model.device.type == DistributedInterface().current_accelerator.type
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/tests_v1/accelerator/test_interface.py
tests_v1/accelerator/test_interface.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import os import pytest import torch.multiprocessing as mp from llamafactory.v1.accelerator.helper import ReduceOp from llamafactory.v1.accelerator.interface import DistributedInterface from llamafactory.v1.utils.env import find_available_port from llamafactory.v1.utils.pytest import dist_env def _all_reduce_tests(local_rank: int, world_size: int, master_port: int): with dist_env(local_rank, world_size, master_port): rank = DistributedInterface().get_rank() world_size = DistributedInterface().get_world_size() assert world_size == 2 y_sum = DistributedInterface().all_reduce(rank + 1.0, op=ReduceOp.SUM) assert y_sum == pytest.approx(3.0) y_mean = DistributedInterface().all_reduce(rank + 1.0, op=ReduceOp.MEAN) assert y_mean == pytest.approx(1.5) y_max = DistributedInterface().all_reduce(rank + 1.0, op=ReduceOp.MAX) assert y_max == pytest.approx(2.0) z = DistributedInterface().all_gather(rank + 1.0) assert z == pytest.approx([1.0, 2.0]) z = DistributedInterface().broadcast(rank + 1.0) assert z == pytest.approx(1.0) def test_all_device(): assert DistributedInterface().get_rank() == int(os.getenv("RANK", "0")) assert DistributedInterface().get_world_size() == int(os.getenv("WORLD_SIZE", "1")) assert DistributedInterface().get_local_rank() == int(os.getenv("LOCAL_RANK", "0")) assert DistributedInterface().get_local_world_size() == int(os.getenv("LOCAL_WORLD_SIZE", "1")) @pytest.mark.runs_on(["cuda", "npu"]) @pytest.mark.require_distributed(2) def test_multi_device(): master_port = find_available_port() mp.spawn(_all_reduce_tests, args=(2, master_port), nprocs=2)
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/tests_v1/core/test_data_loader.py
tests_v1/core/test_data_loader.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Integration tests for DataLoader with different combinations of packing and dynamic batching. Tests the 4 scenarios: a) non pack + non dynamic. b) non pack + dynamic. c) pack + non dynamic. d) pack + dynamic. """ import torch from torch.utils.data import DataLoader as TorchDataLoader from torch.utils.data import Dataset from transformers import AutoTokenizer from llamafactory.v1.config.data_args import DataArguments from llamafactory.v1.core.data_engine import DataEngine from llamafactory.v1.core.trainer_utils.data_collator import ( DefaultCollator, ) from llamafactory.v1.core.trainer_utils.data_loader import DataLoader from llamafactory.v1.plugins.data_plugins.template import QwenTemplate from llamafactory.v1.utils.batching_queue import TextBatchingQueue class TensorDataset(Dataset): """Wrapper dataset that converts DataEngine samples to tensor format.""" def __init__(self, data_engine: DataEngine, processor, template, max_samples: int = None): self.data_engine = data_engine self.processor = processor self.template = template self.max_samples = max_samples or len(data_engine) self.tokenizer = processor.tokenizer if hasattr(processor, "tokenizer") else processor def __len__(self): return min(self.max_samples, len(self.data_engine)) def __getitem__(self, idx): # Get sample from DataEngine sample = self.data_engine[idx] # Extract messages from sample # DataEngine returns samples with format like {"messages": [...], ...} # For llamafactory/v1-sft-demo, the format should have "messages" field messages = None if "messages" in sample: messages = sample["messages"] elif "conversations" in sample: messages = sample["conversations"] elif "conversation" in sample: messages = sample["conversation"] else: # Try to find message-like fields (skip _dataset_name) for key, value in sample.items(): if key.startswith("_"): continue if isinstance(value, list) and len(value) > 0: # Check if it looks like a message list if isinstance(value[0], dict) and "role" in value[0]: messages = value break if messages is None: raise ValueError(f"Could not find messages in sample: {list(sample.keys())}") # Encode messages using template encoded = self.template.encode_messages(self.tokenizer, messages) # Convert to tensors return { "input_ids": torch.tensor(encoded["input_ids"], dtype=torch.long), "attention_mask": torch.tensor(encoded["attention_mask"], dtype=torch.long), "labels": torch.tensor(encoded["labels"], dtype=torch.long), } def create_real_dataset(max_samples: int = 20, batch_size: int = 4): """Create a real dataset using DataEngine.""" data_args = DataArguments(dataset="llamafactory/v1-sft-demo") data_engine = DataEngine(data_args) # Create processor and template processor = AutoTokenizer.from_pretrained("llamafactory/tiny-random-qwen2.5") template = QwenTemplate() # Create tensor dataset raw_data_dataset = TensorDataset(data_engine, processor, template, max_samples=max_samples) # Create torch DataLoader torch_dataloader = TorchDataLoader( raw_data_dataset, batch_size=batch_size, shuffle=False, collate_fn=lambda x: x, ) return torch_dataloader, processor, template class TestDataLoaderNonPackNonDynamic: """Test case a) non pack + non dynamic.""" def test_basic_functionality(self): """Test DataLoader without packing and without dynamic batching.""" # Create real dataset torch_dataloader, processor, template = create_real_dataset(max_samples=80, batch_size=8) # Create collator (non-packing) collator = DefaultCollator(processor=processor, template=template) # Create DataLoader without batching_queue (non-dynamic) data_loader = DataLoader( dataloader=torch_dataloader, collate_fn=collator, num_micro_batch=1, batching_queue=None, ) # Iterate and check results batches = list(iter(data_loader)) assert len(batches) > 0 # Check first batch one_batch = batches[0] micro_batches = one_batch[0] assert "input_ids" in micro_batches assert "attention_mask" in micro_batches assert "labels" in micro_batches assert micro_batches["input_ids"].shape[0] == 1 # batch_size=1 assert micro_batches["input_ids"].ndim == 2 # [batch_size, seq_len] class TestDataLoaderNonPackDynamic: """Test case b) non pack + dynamic.""" def test_basic_functionality(self): """Test DataLoader without packing but with dynamic batching.""" # Create real dataset torch_dataloader, processor, template = create_real_dataset(max_samples=80, batch_size=8) collator = DefaultCollator(processor=processor, template=template) # Create batching queue for dynamic batching batching_queue = TextBatchingQueue( token_micro_bsz=120, buffer_size=8, ) data_loader = DataLoader( dataloader=torch_dataloader, collate_fn=collator, num_micro_batch=4, batching_queue=batching_queue, ) # Iterate and check batches = list(iter(data_loader)) micro_batch_tokens_first = [micro_batch["attention_mask"].sum() for micro_batch in batches[0]] assert all(num_tokens <= 120 for num_tokens in micro_batch_tokens_first) assert len(batches) > 0
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/tests_v1/core/test_model_loader.py
tests_v1/core/test_model_loader.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from llamafactory.v1.config.model_args import ModelArguments, PluginConfig from llamafactory.v1.core.model_loader import ModelLoader def test_tiny_qwen(): from transformers import Qwen2Config, Qwen2ForCausalLM, Qwen2TokenizerFast model_args = ModelArguments(model="llamafactory/tiny-random-qwen2.5") model_loader = ModelLoader(model_args) assert isinstance(model_loader.processor, Qwen2TokenizerFast) assert isinstance(model_loader.model.config, Qwen2Config) assert isinstance(model_loader.model, Qwen2ForCausalLM) assert model_loader.model.dtype == torch.bfloat16 def test_tiny_qwen_with_kernel_plugin(): from transformers import Qwen2ForCausalLM from llamafactory.v1.plugins.model_plugins.kernels.ops.rms_norm.npu_rms_norm import npu_rms_norm_forward model_args = ModelArguments( model="llamafactory/tiny-random-qwen2.5", kernel_config=PluginConfig(name="auto", include_kernels="auto") ) model_loader = ModelLoader(model_args) # test enable apply kernel plugin if hasattr(torch, "npu"): assert model_loader.model.model.layers[0].input_layernorm.forward.__code__ == npu_rms_norm_forward.__code__ else: assert model_loader.model.model.layers[0].input_layernorm.forward.__code__ != npu_rms_norm_forward.__code__ assert isinstance(model_loader.model, Qwen2ForCausalLM) if __name__ == "__main__": test_tiny_qwen() test_tiny_qwen_with_kernel_plugin()
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/tests_v1/core/test_data_engine.py
tests_v1/core/test_data_engine.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import random import pytest from datasets import load_dataset from llamafactory.v1.config.data_args import DataArguments from llamafactory.v1.core.data_engine import DataEngine @pytest.mark.parametrize("num_samples", [16]) def test_map_dataset(num_samples: int): data_args = DataArguments(dataset="llamafactory/v1-sft-demo") data_engine = DataEngine(data_args) original_data = load_dataset("llamafactory/v1-sft-demo", split="train") indexes = random.choices(range(len(data_engine)), k=num_samples) for index in indexes: print(data_engine[index]) assert data_engine[index] == {"_dataset_name": "default", **original_data[index]} if __name__ == "__main__": test_map_dataset(1)
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/tests_v1/config/test_args_parser.py
tests_v1/config/test_args_parser.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import pathlib import sys from unittest.mock import patch from llamafactory.v1.config.arg_parser import get_args def test_get_args_from_yaml(tmp_path: pathlib.Path): config_yaml = """ ### model model: "llamafactory/tiny-random-qwen2.5" trust_remote_code: true model_class: "llm" kernel_config: name: "auto" include_kernels: "auto" # choice: null/true/false/auto/kernel_id1,kernel_id2,kernel_id3, default is null peft_config: name: "lora" lora_rank: 0.8 quant_config: null ### data dataset: "llamafactory/tiny-supervised-dataset" cutoff_len: 2048 ### training output_dir: "outputs/test_run" micro_batch_size: 1 global_batch_size: 1 learning_rate: 1.0e-4 bf16: false dist_config: null ### sample sample_backend: "hf" max_new_tokens: 128 """ config_file = tmp_path / "config.yaml" config_file.write_text(config_yaml, encoding="utf-8") test_argv = ["test_args_parser.py", str(config_file)] with patch.object(sys, "argv", test_argv): data_args, model_args, training_args, sample_args = get_args() assert training_args.output_dir == "outputs/test_run" assert training_args.micro_batch_size == 1 assert training_args.global_batch_size == 1 assert training_args.learning_rate == 1.0e-4 assert training_args.bf16 is False assert training_args.dist_config is None assert model_args.model == "llamafactory/tiny-random-qwen2.5" assert model_args.kernel_config.name == "auto" assert model_args.kernel_config.get("include_kernels") == "auto" assert model_args.peft_config.name == "lora" assert model_args.peft_config.get("lora_rank") == 0.8
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/train.py
src/train.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from llamafactory.train.tuner import run_exp def main(): run_exp() def _mp_fn(index): # For xla_spawn (TPUs) run_exp() if __name__ == "__main__": main()
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/api.py
src/api.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import os import uvicorn from llamafactory.api.app import create_app from llamafactory.chat import ChatModel def main(): chat_model = ChatModel() app = create_app(chat_model) api_host = os.getenv("API_HOST", "0.0.0.0") api_port = int(os.getenv("API_PORT", "8000")) print(f"Visit http://localhost:{api_port}/docs for API document.") uvicorn.run(app, host=api_host, port=api_port) if __name__ == "__main__": main()
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/webui.py
src/webui.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import os from llamafactory.extras.misc import fix_proxy, is_env_enabled from llamafactory.webui.interface import create_ui def main(): gradio_ipv6 = is_env_enabled("GRADIO_IPV6") gradio_share = is_env_enabled("GRADIO_SHARE") server_name = os.getenv("GRADIO_SERVER_NAME", "[::]" if gradio_ipv6 else "0.0.0.0") print("Visit http://ip:port for Web UI, e.g., http://127.0.0.1:7860") fix_proxy(ipv6_enabled=gradio_ipv6) create_ui().queue().launch(share=gradio_share, server_name=server_name, inbrowser=True) if __name__ == "__main__": main()
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/cli.py
src/llamafactory/cli.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. def main(): from .extras.misc import is_env_enabled if is_env_enabled("USE_V1"): from .v1 import launcher else: from . import launcher launcher.launch() if __name__ == "__main__": from multiprocessing import freeze_support freeze_support() main()
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/launcher.py
src/llamafactory/launcher.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import os import subprocess import sys from copy import deepcopy USAGE = ( "-" * 70 + "\n" + "| Usage: |\n" + "| llamafactory-cli api -h: launch an OpenAI-style API server |\n" + "| llamafactory-cli chat -h: launch a chat interface in CLI |\n" + "| llamafactory-cli export -h: merge LoRA adapters and export model |\n" + "| llamafactory-cli train -h: train models |\n" + "| llamafactory-cli webchat -h: launch a chat interface in Web UI |\n" + "| llamafactory-cli webui: launch LlamaBoard |\n" + "| llamafactory-cli env: show environment info |\n" + "| llamafactory-cli version: show version info |\n" + "| Hint: You can use `lmf` as a shortcut for `llamafactory-cli`. |\n" + "-" * 70 ) def launch(): from .extras import logging from .extras.env import VERSION, print_env from .extras.misc import find_available_port, get_device_count, is_env_enabled, use_kt, use_ray logger = logging.get_logger(__name__) WELCOME = ( "-" * 58 + "\n" + f"| Welcome to LLaMA Factory, version {VERSION}" + " " * (21 - len(VERSION)) + "|\n|" + " " * 56 + "|\n" + "| Project page: https://github.com/hiyouga/LLaMA-Factory |\n" + "-" * 58 ) command = sys.argv.pop(1) if len(sys.argv) > 1 else "help" if is_env_enabled("USE_MCA"): # force use torchrun os.environ["FORCE_TORCHRUN"] = "1" if command == "train" and ( is_env_enabled("FORCE_TORCHRUN") or (get_device_count() > 1 and not use_ray() and not use_kt()) ): # launch distributed training nnodes = os.getenv("NNODES", "1") node_rank = os.getenv("NODE_RANK", "0") nproc_per_node = os.getenv("NPROC_PER_NODE", str(get_device_count())) master_addr = os.getenv("MASTER_ADDR", "127.0.0.1") master_port = os.getenv("MASTER_PORT", str(find_available_port())) logger.info_rank0(f"Initializing {nproc_per_node} distributed tasks at: {master_addr}:{master_port}") if int(nnodes) > 1: logger.info_rank0(f"Multi-node training enabled: num nodes: {nnodes}, node rank: {node_rank}") # elastic launch support max_restarts = os.getenv("MAX_RESTARTS", "0") rdzv_id = os.getenv("RDZV_ID") min_nnodes = os.getenv("MIN_NNODES") max_nnodes = os.getenv("MAX_NNODES") env = deepcopy(os.environ) if is_env_enabled("OPTIM_TORCH", "1"): # optimize DDP, see https://zhuanlan.zhihu.com/p/671834539 env["PYTORCH_CUDA_ALLOC_CONF"] = "expandable_segments:True" env["TORCH_NCCL_AVOID_RECORD_STREAMS"] = "1" if rdzv_id is not None: # launch elastic job with fault tolerant support when possible # see also https://docs.pytorch.org/docs/stable/elastic/train_script.html rdzv_nnodes = nnodes # elastic number of nodes if MIN_NNODES and MAX_NNODES are set if min_nnodes is not None and max_nnodes is not None: rdzv_nnodes = f"{min_nnodes}:{max_nnodes}" process = subprocess.run( ( "torchrun --nnodes {rdzv_nnodes} --nproc-per-node {nproc_per_node} " "--rdzv-id {rdzv_id} --rdzv-backend c10d --rdzv-endpoint {master_addr}:{master_port} " "--max-restarts {max_restarts} {file_name} {args}" ) .format( rdzv_nnodes=rdzv_nnodes, nproc_per_node=nproc_per_node, rdzv_id=rdzv_id, master_addr=master_addr, master_port=master_port, max_restarts=max_restarts, file_name=__file__, args=" ".join(sys.argv[1:]), ) .split(), env=env, check=True, ) else: # NOTE: DO NOT USE shell=True to avoid security risk process = subprocess.run( ( "torchrun --nnodes {nnodes} --node_rank {node_rank} --nproc_per_node {nproc_per_node} " "--master_addr {master_addr} --master_port {master_port} {file_name} {args}" ) .format( nnodes=nnodes, node_rank=node_rank, nproc_per_node=nproc_per_node, master_addr=master_addr, master_port=master_port, file_name=__file__, args=" ".join(sys.argv[1:]), ) .split(), env=env, check=True, ) sys.exit(process.returncode) elif command == "api": from .api.app import run_api run_api() elif command == "chat": from .chat.chat_model import run_chat run_chat() elif command == "eval": raise NotImplementedError("Evaluation will be deprecated in the future.") elif command == "export": from .train.tuner import export_model export_model() elif command == "train": from .train.tuner import run_exp run_exp() elif command == "webchat": from .webui.interface import run_web_demo run_web_demo() elif command == "webui": from .webui.interface import run_web_ui run_web_ui() elif command == "env": print_env() elif command == "version": print(WELCOME) elif command == "help": print(USAGE) else: print(f"Unknown command: {command}.\n{USAGE}") if __name__ == "__main__": from llamafactory.train.tuner import run_exp # use absolute import run_exp()
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/__init__.py
src/llamafactory/__init__.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. r"""Efficient fine-tuning of large language models. Level: api, webui > chat, eval, train > data, model > hparams > extras Disable version checking: DISABLE_VERSION_CHECK=1 Enable VRAM recording: RECORD_VRAM=1 Force using torchrun: FORCE_TORCHRUN=1 Set logging verbosity: LLAMAFACTORY_VERBOSITY=WARN Use modelscope: USE_MODELSCOPE_HUB=1 Use openmind: USE_OPENMIND_HUB=1 """ from .extras.env import VERSION __version__ = VERSION
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/third_party/__init__.py
src/llamafactory/third_party/__init__.py
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/third_party/muon/muon.py
src/llamafactory/third_party/muon/muon.py
# Copyright 2025 Moonshot AI and the LlamaFactory team. # # This code is based on the MoonshotAI's Moonlight library. # https://github.com/MoonshotAI/Moonlight/blob/master/examples/toy_train.py # and the Keller Jordan's Muon library. # https://github.com/KellerJordan/Muon/blob/master/muon.py # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # # MIT License # # Copyright (c) 2025 Moonshot AI # Copyright (c) 2024 Keller Jordan # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitation the rights # to use, copy, modify, merge, publish, distribute, sublicense, and/or sell # copies of the Software, and to permit persons to whom the Software is # furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in all # copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE # SOFTWARE. import math import torch def zeropower_via_newtonschulz5(G: "torch.Tensor", steps: int) -> "torch.Tensor": """Newton-Schulz iteration to compute the zeroth power / orthogonalization of G. We opt to use a quintic iteration whose coefficients are selected to maximize the slope at zero. For the purpose of minimizing steps, it turns out to be empirically effective to keep increasing the slope at zero even beyond the point where the iteration no longer converges all the way to one everywhere on the interval. This iteration therefore does not produce UV^T but rather something like US'V^T where S' is diagonal with S_{ii}' ~ Uniform(0.5, 1.5), which turns out not to hurt model performance at all relative to UV^T, where USV^T = G is the SVD. """ assert len(G.shape) == 2 a, b, c = (3.4445, -4.7750, 2.0315) X = G.bfloat16() if G.size(0) > G.size(1): X = X.T # Ensure spectral norm is at most 1 X = X / (X.norm() + 1e-7) # Perform the NS iterations for _ in range(steps): A = X @ X.T B = b * A + c * A @ A # adapted from suggestion by @jxbz, @leloykun, and @YouJiacheng X = a * X + B @ X if G.size(0) > G.size(1): X = X.T return X class Muon(torch.optim.Optimizer): """Muon - MomentUm Orthogonalized by Newton-schulz. Muon internally runs standard SGD-momentum, and then performs an orthogonalization post- processing step, in which each 2D parameter's update is replaced with the nearest orthogonal matrix. To efficiently orthogonalize each update, we use a Newton-Schulz iteration, which has the advantage that it can be stably run in bfloat16 on the GPU. Some warnings: - We believe this optimizer is unlikely to work well for training with small batch size. - We believe it may not work well for finetuning pretrained models, but we haven't tested this. Arguments: muon_params: The parameters to be optimized by Muon. lr: The learning rate. The updates will have spectral norm of `lr`. (0.02 is a good default) momentum: The momentum used by the internal SGD. (0.95 is a good default) nesterov: Whether to use Nesterov-style momentum in the internal SGD. (recommended) ns_steps: The number of Newton-Schulz iterations to run. (6 is probably always enough) adamw_params: The parameters to be optimized by AdamW. Any parameters in `muon_params` which are {0, 1}-D or are detected as being the embed or lm_head will be optimized by AdamW as well. adamw_lr: The learning rate for the internal AdamW. adamw_betas: The betas for the internal AdamW. adamw_eps: The epsilon for the internal AdamW. adamw_wd: The weight decay for the internal AdamW. """ def __init__( self, lr=1e-3, wd=0.1, muon_params=None, momentum=0.95, nesterov=True, ns_steps=5, adamw_params=None, adamw_betas=(0.9, 0.95), adamw_eps=1e-8, ): defaults = dict( lr=lr, wd=wd, momentum=momentum, nesterov=nesterov, ns_steps=ns_steps, adamw_betas=adamw_betas, adamw_eps=adamw_eps, ) params = list(muon_params) adamw_params = list(adamw_params) if adamw_params is not None else [] params.extend(adamw_params) super().__init__(params, defaults) # Sort parameters into those for which we will use Muon, and those for which we will not for p in muon_params: # Use Muon for every parameter in muon_params which is >= 2D and doesn't look like an embedding or head layer assert p.ndim == 2, p.ndim self.state[p]["use_muon"] = True for p in adamw_params: # Do not use Muon for parameters in adamw_params self.state[p]["use_muon"] = False def adjust_lr_for_muon(self, lr: float, param_shape: list[int]) -> float: A, B = param_shape[:2] # We adjust the learning rate and weight decay based on the size of the parameter matrix # as describted in the paper adjusted_ratio = 0.2 * math.sqrt(max(A, B)) adjusted_lr = lr * adjusted_ratio return adjusted_lr def step(self, closure=None): """Perform a single optimization step. Args: closure (Callable, optional): A closure that reevaluates the model and returns the loss. """ loss = None if closure is not None: with torch.enable_grad(): loss = closure() for group in self.param_groups: # Muon loop params = [p for p in group["params"] if self.state[p]["use_muon"]] lr = group["lr"] wd = group["wd"] momentum = group["momentum"] # generate weight updates in distributed fashion for p in params: # sanity check g = p.grad if g is None: continue if g.ndim > 2: g = g.view(g.size(0), -1) assert g is not None # calc update state = self.state[p] if "momentum_buffer" not in state: state["momentum_buffer"] = torch.zeros_like(g) buf = state["momentum_buffer"] buf.mul_(momentum).add_(g) if group["nesterov"]: g = g.add(buf, alpha=momentum) else: g = buf u = zeropower_via_newtonschulz5(g, steps=group["ns_steps"]) # scale update adjusted_lr = self.adjust_lr_for_muon(lr, p.shape) # apply weight decay p.data.mul_(1 - lr * wd) # apply update p.data.add_(u, alpha=-adjusted_lr) # Adam backup params = [p for p in group["params"] if not self.state[p]["use_muon"]] lr = group["lr"] beta1, beta2 = group["adamw_betas"] eps = group["adamw_eps"] weight_decay = group["wd"] for p in params: g = p.grad if g is None: continue state = self.state[p] if "step" not in state: state["step"] = 0 state["moment1"] = torch.zeros_like(g) state["moment2"] = torch.zeros_like(g) state["step"] += 1 step = state["step"] buf1 = state["moment1"] buf2 = state["moment2"] buf1.lerp_(g, 1 - beta1) buf2.lerp_(g.square(), 1 - beta2) g = buf1 / (eps + buf2.sqrt()) bias_correction1 = 1 - beta1**step bias_correction2 = 1 - beta2**step scale = bias_correction1 / bias_correction2**0.5 p.data.mul_(1 - lr * weight_decay) p.data.add_(g, alpha=-lr / scale) return loss
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/third_party/muon/__init__.py
src/llamafactory/third_party/muon/__init__.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from .muon import Muon __all__ = ["Muon"]
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/chat/kt_engine.py
src/llamafactory/chat/kt_engine.py
# Copyright 2025 the KVCache.AI team, Approaching AI, and the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import asyncio import os import platform from collections.abc import AsyncGenerator from threading import Thread from typing import TYPE_CHECKING, Any, Optional import torch from typing_extensions import override from ..data import get_template_and_fix_tokenizer from ..extras import logging from ..extras.constants import EngineName from ..model import load_model, load_tokenizer from .base_engine import BaseEngine, Response if TYPE_CHECKING: from transformers import PreTrainedTokenizer from trl import PreTrainedModelWrapper from ..data.mm_plugin import AudioInput, ImageInput, VideoInput from ..hparams import DataArguments, FinetuningArguments, GeneratingArguments, ModelArguments from ktransformers.operators.flashinfer_wrapper import flashinfer_enabled from ktransformers.server.config.config import Config from ktransformers.util.utils import ( get_compute_capability, prefill_and_generate_capture, ) from ktransformers.util.vendors import GPUVendor, device_manager logger = logging.get_logger(__name__) class KTransformersEngine(BaseEngine): def __init__( self, model_args: "ModelArguments", data_args: "DataArguments", finetuning_args: "FinetuningArguments", generating_args: "GeneratingArguments", ) -> None: self.name = EngineName.KT self.can_generate = finetuning_args.stage == "sft" tok_mod = load_tokenizer(model_args) self.tokenizer = tok_mod["tokenizer"] self.tokenizer.padding_side = "left" if self.can_generate else "right" self.template = get_template_and_fix_tokenizer(self.tokenizer, data_args) self.model = load_model( self.tokenizer, model_args, finetuning_args, is_trainable=False, add_valuehead=(not self.can_generate) ) self.generating_args = generating_args.to_dict() self.max_new_tokens = model_args.kt_maxlen self.use_cuda_graph = model_args.kt_use_cuda_graph self.mode = model_args.kt_mode self.force_think = model_args.kt_force_think self.chunk_size = model_args.chunk_size try: asyncio.get_event_loop() except RuntimeError: loop = asyncio.new_event_loop() asyncio.set_event_loop(loop) self.semaphore = asyncio.Semaphore(int(os.getenv("MAX_CONCURRENT", "1"))) @staticmethod @torch.inference_mode() def _get_scores( model: "PreTrainedModelWrapper", tokenizer: "PreTrainedTokenizer", batch_input: list[str], input_kwargs: Optional[dict[str, Any]] = {}, ) -> list[float]: max_length: Optional[int] = input_kwargs.pop("max_length", None) device = getattr(model.pretrained_model, "device", "cuda") inputs = tokenizer( batch_input, padding=True, truncation=True, max_length=max_length or getattr(model.config, "max_position_embeddings", 1024), return_tensors="pt", add_special_tokens=False, ).to(device) values: torch.Tensor = model(**inputs, return_dict=True, use_cache=False)[-1] scores = values.gather(dim=-1, index=(inputs["attention_mask"].sum(dim=-1, keepdim=True) - 1)) return scores async def _generate( self, messages: list[dict[str, str]], system: Optional[str] = None, tools: Optional[str] = None, **input_kwargs, ) -> AsyncGenerator[str, None]: paired = messages + [{"role": "assistant", "content": ""}] prompt_ids, _ = self.template.encode_oneturn(self.tokenizer, paired, system, tools) prompt_len = len(prompt_ids) max_length: Optional[int] = input_kwargs.pop("max_length", None) max_new_tokens: Optional[int] = input_kwargs.pop("max_new_tokens", None) if "max_new_tokens" in self.generating_args: max_tokens = int(self.generating_args["max_new_tokens"]) elif "max_length" in self.generating_args: gl = int(self.generating_args["max_length"]) max_tokens = gl - prompt_len if gl > prompt_len else 1 else: max_tokens = self.max_new_tokens or 256 if max_length is not None: max_tokens = max(max_length - prompt_len, 1) if max_new_tokens is not None: max_tokens = int(max_new_tokens) max_tokens = max(1, int(max_tokens)) if self.mode == "long_context": max_len_cfg = Config().long_context_config["max_seq_len"] need = prompt_len + max_tokens assert max_len_cfg > need, f"please set max_seq_len > {need} in ~/.ktransformers/config.yaml" device = next(self.model.parameters()).device input_tensor = torch.tensor([prompt_ids], dtype=torch.long, device=device) if self.force_think: think = torch.tensor( [self.tokenizer.encode("<think>\n", add_special_tokens=False)], dtype=torch.long, device=device ) input_tensor = torch.cat([input_tensor, think], dim=1) use_flashinfer = ( platform.system() != "Windows" and getattr(self.model.config, "architectures", [""])[0] in {"DeepseekV2ForCausalLM", "DeepseekV3ForCausalLM"} and flashinfer_enabled and get_compute_capability() >= 8 and device_manager.gpu_vendor == GPUVendor.NVIDIA ) def make_gen(): if use_flashinfer: return prefill_and_generate_capture( self.model, self.tokenizer, input_tensor, max_tokens, self.use_cuda_graph, mode=self.mode, force_think=self.force_think, chunk_size=self.chunk_size, use_flashinfer_mla=True, num_heads=self.model.config.num_attention_heads, head_dim_ckv=getattr(self.model.config, "kv_lora_rank", 0), head_dim_kpe=getattr(self.model.config, "qk_rope_head_dim", 0), q_head_dim=getattr(self.model.config, "qk_rope_head_dim", 0) + getattr(self.model.config, "qk_nope_head_dim", 0), echo_stream=False, ) else: return prefill_and_generate_capture( self.model, self.tokenizer, input_tensor, max_tokens, self.use_cuda_graph, mode=self.mode, force_think=self.force_think, chunk_size=self.chunk_size, echo_stream=False, ) loop = asyncio.get_running_loop() q: asyncio.Queue[Optional[str]] = asyncio.Queue() def producer(): try: gen = make_gen() if hasattr(gen, "__aiter__"): async def drain_async(): async for t in gen: loop.call_soon_threadsafe(q.put_nowait, t if isinstance(t, str) else str(t)) asyncio.run(drain_async()) elif hasattr(gen, "__iter__"): for t in gen: loop.call_soon_threadsafe(q.put_nowait, t if isinstance(t, str) else str(t)) else: loop.call_soon_threadsafe(q.put_nowait, gen if isinstance(gen, str) else str(gen)) finally: loop.call_soon_threadsafe(q.put_nowait, None) Thread(target=producer, daemon=True).start() while True: item = await q.get() if item is None: break yield item @override async def chat( self, messages: list[dict[str, str]], system: Optional[str] = None, tools: Optional[str] = None, images: Optional[list["ImageInput"]] = None, videos: Optional[list["VideoInput"]] = None, audios: Optional[list["AudioInput"]] = None, **input_kwargs, ) -> list["Response"]: if not self.can_generate: raise ValueError("The current model does not support `chat`.") async with self.semaphore: produced = "" final_text = "" async for t in self._generate(messages, system, tools, **input_kwargs): delta = t produced = produced + delta if delta: final_text += delta prompt_ids, _ = self.template.encode_oneturn( self.tokenizer, messages + [{"role": "assistant", "content": ""}], system, tools ) return [ Response( response_text=final_text, response_length=len(self.tokenizer.encode(final_text, add_special_tokens=False)), prompt_length=len(prompt_ids), finish_reason="stop", ) ] @override async def stream_chat( self, messages: list[dict[str, str]], system: Optional[str] = None, tools: Optional[str] = None, images: Optional[list["ImageInput"]] = None, videos: Optional[list["VideoInput"]] = None, audios: Optional[list["AudioInput"]] = None, **input_kwargs, ) -> AsyncGenerator[str, None]: if not self.can_generate: raise ValueError("The current model does not support `stream_chat`.") async with self.semaphore: produced = "" async for t in self._generate(messages, system, tools, **input_kwargs): delta = t[len(produced) :] if t.startswith(produced) else t produced = t if delta: yield delta @override async def get_scores( self, batch_input: list[str], **input_kwargs, ) -> list[float]: if self.can_generate: raise ValueError("Cannot get scores using an auto-regressive model.") args = (self.model, self.tokenizer, batch_input, input_kwargs) async with self.semaphore: return await asyncio.to_thread(self._get_scores, *args)
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/chat/hf_engine.py
src/llamafactory/chat/hf_engine.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import asyncio import os from collections.abc import AsyncGenerator, Callable from threading import Thread from typing import TYPE_CHECKING, Any, Optional, Union import torch from transformers import GenerationConfig, TextIteratorStreamer from typing_extensions import override from ..data import get_template_and_fix_tokenizer from ..extras import logging from ..extras.constants import AUDIO_PLACEHOLDER, IMAGE_PLACEHOLDER, VIDEO_PLACEHOLDER, EngineName from ..model import load_model, load_tokenizer from .base_engine import BaseEngine, Response if TYPE_CHECKING: from transformers import PreTrainedModel, PreTrainedTokenizer, ProcessorMixin from trl import PreTrainedModelWrapper from ..data import Template from ..data.mm_plugin import AudioInput, ImageInput, VideoInput from ..hparams import DataArguments, FinetuningArguments, GeneratingArguments, ModelArguments logger = logging.get_logger(__name__) class HuggingfaceEngine(BaseEngine): def __init__( self, model_args: "ModelArguments", data_args: "DataArguments", finetuning_args: "FinetuningArguments", generating_args: "GeneratingArguments", ) -> None: self.name = EngineName.HF self.can_generate = finetuning_args.stage == "sft" tokenizer_module = load_tokenizer(model_args) self.tokenizer = tokenizer_module["tokenizer"] self.processor = tokenizer_module["processor"] self.tokenizer.padding_side = "left" if self.can_generate else "right" self.template = get_template_and_fix_tokenizer(self.tokenizer, data_args) self.model = load_model( self.tokenizer, model_args, finetuning_args, is_trainable=False, add_valuehead=(not self.can_generate) ) # must after fixing tokenizer to resize vocab self.generating_args = generating_args.to_dict() try: asyncio.get_event_loop() except RuntimeError: logger.warning_rank0_once("There is no current event loop, creating a new one.") loop = asyncio.new_event_loop() asyncio.set_event_loop(loop) self.semaphore = asyncio.Semaphore(int(os.getenv("MAX_CONCURRENT", "1"))) @staticmethod def _process_args( model: "PreTrainedModel", tokenizer: "PreTrainedTokenizer", processor: Optional["ProcessorMixin"], template: "Template", generating_args: dict[str, Any], messages: list[dict[str, str]], system: Optional[str] = None, tools: Optional[str] = None, images: Optional[list["ImageInput"]] = None, videos: Optional[list["VideoInput"]] = None, audios: Optional[list["AudioInput"]] = None, input_kwargs: Optional[dict[str, Any]] = {}, ) -> tuple[dict[str, Any], int]: mm_input_dict = {"images": [], "videos": [], "audios": [], "imglens": [0], "vidlens": [0], "audlens": [0]} if images is not None: mm_input_dict.update({"images": images, "imglens": [len(images)]}) if not any(IMAGE_PLACEHOLDER in message["content"] for message in messages): messages[0]["content"] = IMAGE_PLACEHOLDER * len(images) + messages[0]["content"] if videos is not None: mm_input_dict.update({"videos": videos, "vidlens": [len(videos)]}) if not any(VIDEO_PLACEHOLDER in message["content"] for message in messages): messages[0]["content"] = VIDEO_PLACEHOLDER * len(videos) + messages[0]["content"] if audios is not None: mm_input_dict.update({"audios": audios, "audlens": [len(audios)]}) if not any(AUDIO_PLACEHOLDER in message["content"] for message in messages): messages[0]["content"] = AUDIO_PLACEHOLDER * len(audios) + messages[0]["content"] messages = template.mm_plugin.process_messages( messages, mm_input_dict["images"], mm_input_dict["videos"], mm_input_dict["audios"], processor ) paired_messages = messages + [{"role": "assistant", "content": ""}] prompt_ids, _ = template.encode_oneturn(tokenizer, paired_messages, system, tools) prompt_ids, _ = template.mm_plugin.process_token_ids( prompt_ids, None, mm_input_dict["images"], mm_input_dict["videos"], mm_input_dict["audios"], tokenizer, processor, ) prompt_length = len(prompt_ids) inputs = torch.tensor([prompt_ids], device=model.device) attention_mask = torch.ones_like(inputs, dtype=torch.long) do_sample: Optional[bool] = input_kwargs.pop("do_sample", None) temperature: Optional[float] = input_kwargs.pop("temperature", None) top_p: Optional[float] = input_kwargs.pop("top_p", None) top_k: Optional[float] = input_kwargs.pop("top_k", None) num_return_sequences: int = input_kwargs.pop("num_return_sequences", 1) repetition_penalty: Optional[float] = input_kwargs.pop("repetition_penalty", None) length_penalty: Optional[float] = input_kwargs.pop("length_penalty", None) skip_special_tokens: Optional[bool] = input_kwargs.pop("skip_special_tokens", None) max_length: Optional[int] = input_kwargs.pop("max_length", None) max_new_tokens: Optional[int] = input_kwargs.pop("max_new_tokens", None) stop: Optional[Union[str, list[str]]] = input_kwargs.pop("stop", None) if stop is not None: logger.warning_rank0("Stop parameter is not supported by the huggingface engine yet.") generating_args = generating_args.copy() generating_args.update( dict( do_sample=do_sample if do_sample is not None else generating_args["do_sample"], temperature=temperature if temperature is not None else generating_args["temperature"], top_p=top_p if top_p is not None else generating_args["top_p"], top_k=top_k if top_k is not None else generating_args["top_k"], num_return_sequences=num_return_sequences, repetition_penalty=repetition_penalty if repetition_penalty is not None else generating_args["repetition_penalty"], length_penalty=length_penalty if length_penalty is not None else generating_args["length_penalty"], skip_special_tokens=skip_special_tokens if skip_special_tokens is not None else generating_args["skip_special_tokens"], eos_token_id=template.get_stop_token_ids(tokenizer), pad_token_id=tokenizer.pad_token_id, ) ) if isinstance(num_return_sequences, int) and num_return_sequences > 1: # do_sample needs temperature > 0 generating_args["do_sample"] = True generating_args["temperature"] = generating_args["temperature"] or 1.0 if not generating_args["temperature"]: generating_args["do_sample"] = False if not generating_args["do_sample"]: generating_args.pop("temperature", None) generating_args.pop("top_p", None) if max_length: generating_args.pop("max_new_tokens", None) generating_args["max_length"] = max_length if max_new_tokens: generating_args.pop("max_length", None) generating_args["max_new_tokens"] = max_new_tokens gen_kwargs = dict( inputs=inputs, attention_mask=attention_mask, generation_config=GenerationConfig(**generating_args), ) mm_inputs = template.mm_plugin.get_mm_inputs(**mm_input_dict, batch_ids=[prompt_ids], processor=processor) for key, value in mm_inputs.items(): if isinstance(value, list) and isinstance(value[0], torch.Tensor): # for pixtral inputs value = torch.stack(value) # assume they have same sizes elif ( isinstance(value, list) and isinstance(value[0], list) and isinstance(value[0][0], torch.Tensor) ): # for minicpmv inputs value = torch.stack([torch.stack(v) for v in value]) elif not isinstance(value, torch.Tensor): value = torch.tensor(value) if torch.is_floating_point(value): # cast data dtype for paligemma value = value.to(model.dtype) if key == "second_per_grid_ts": # qwen2.5vl special case gen_kwargs[key] = value.tolist() else: gen_kwargs[key] = value.to(model.device) if getattr(model.config, "model_type", None) in ["minicpmv", "minicpmo"]: gen_kwargs["input_ids"] = inputs gen_kwargs["tokenizer"] = tokenizer if "audio_feature_lens" in mm_inputs: gen_kwargs["audio_feature_lens"] = mm_inputs["audio_feature_lens"] gen_kwargs.pop("image_sizes", None) return gen_kwargs, prompt_length @staticmethod @torch.inference_mode() def _chat( model: "PreTrainedModel", tokenizer: "PreTrainedTokenizer", processor: Optional["ProcessorMixin"], template: "Template", generating_args: dict[str, Any], messages: list[dict[str, str]], system: Optional[str] = None, tools: Optional[str] = None, images: Optional[list["ImageInput"]] = None, videos: Optional[list["VideoInput"]] = None, audios: Optional[list["AudioInput"]] = None, input_kwargs: Optional[dict[str, Any]] = {}, ) -> list["Response"]: gen_kwargs, prompt_length = HuggingfaceEngine._process_args( model, tokenizer, processor, template, generating_args, messages, system, tools, images, videos, audios, input_kwargs, ) generate_output = model.generate(**gen_kwargs) if isinstance(generate_output, tuple): generate_output = generate_output[1][0] # post-process the minicpm_o output response_ids = generate_output[:, prompt_length:] response = tokenizer.batch_decode( response_ids, skip_special_tokens=getattr(gen_kwargs["generation_config"], "skip_special_tokens", True), clean_up_tokenization_spaces=True, ) results = [] for i in range(len(response)): eos_index = (response_ids[i] == tokenizer.eos_token_id).nonzero() response_length = (eos_index[0].item() + 1) if len(eos_index) else len(response_ids[i]) results.append( Response( response_text=response[i], response_length=response_length, prompt_length=prompt_length, finish_reason="stop" if len(eos_index) else "length", ) ) return results @staticmethod @torch.inference_mode() def _stream_chat( model: "PreTrainedModel", tokenizer: "PreTrainedTokenizer", processor: Optional["ProcessorMixin"], template: "Template", generating_args: dict[str, Any], messages: list[dict[str, str]], system: Optional[str] = None, tools: Optional[str] = None, images: Optional[list["ImageInput"]] = None, videos: Optional[list["VideoInput"]] = None, audios: Optional[list["AudioInput"]] = None, input_kwargs: Optional[dict[str, Any]] = {}, ) -> Callable[[], str]: gen_kwargs, _ = HuggingfaceEngine._process_args( model, tokenizer, processor, template, generating_args, messages, system, tools, images, videos, audios, input_kwargs, ) streamer = TextIteratorStreamer( tokenizer, skip_prompt=True, skip_special_tokens=getattr(gen_kwargs["generation_config"], "skip_special_tokens", True), ) gen_kwargs["streamer"] = streamer thread = Thread(target=model.generate, kwargs=gen_kwargs, daemon=True) thread.start() def stream(): try: return streamer.__next__() except StopIteration: raise StopAsyncIteration() return stream @staticmethod @torch.inference_mode() def _get_scores( model: "PreTrainedModelWrapper", tokenizer: "PreTrainedTokenizer", batch_input: list[str], input_kwargs: Optional[dict[str, Any]] = {}, ) -> list[float]: max_length: Optional[int] = input_kwargs.pop("max_length", None) device = getattr(model.pretrained_model, "device", "cuda") inputs: dict[str, torch.Tensor] = tokenizer( batch_input, padding=True, truncation=True, max_length=max_length or getattr(model.config, "max_position_embeddings", 1024), return_tensors="pt", add_special_tokens=False, ).to(device) values: torch.Tensor = model(**inputs, return_dict=True, use_cache=False)[-1] scores = values.gather(dim=-1, index=(inputs["attention_mask"].sum(dim=-1, keepdim=True) - 1)) return scores @override async def chat( self, messages: list[dict[str, str]], system: Optional[str] = None, tools: Optional[str] = None, images: Optional[list["ImageInput"]] = None, videos: Optional[list["VideoInput"]] = None, audios: Optional[list["AudioInput"]] = None, **input_kwargs, ) -> list["Response"]: if not self.can_generate: raise ValueError("The current model does not support `chat`.") input_args = ( self.model, self.tokenizer, self.processor, self.template, self.generating_args, messages, system, tools, images, videos, audios, input_kwargs, ) async with self.semaphore: return await asyncio.to_thread(self._chat, *input_args) @override async def stream_chat( self, messages: list[dict[str, str]], system: Optional[str] = None, tools: Optional[str] = None, images: Optional[list["ImageInput"]] = None, videos: Optional[list["VideoInput"]] = None, audios: Optional[list["AudioInput"]] = None, **input_kwargs, ) -> AsyncGenerator[str, None]: if not self.can_generate: raise ValueError("The current model does not support `stream_chat`.") input_args = ( self.model, self.tokenizer, self.processor, self.template, self.generating_args, messages, system, tools, images, videos, audios, input_kwargs, ) async with self.semaphore: stream = self._stream_chat(*input_args) while True: try: yield await asyncio.to_thread(stream) except StopAsyncIteration: break @override async def get_scores( self, batch_input: list[str], **input_kwargs, ) -> list[float]: if self.can_generate: raise ValueError("Cannot get scores using an auto-regressive model.") input_args = (self.model, self.tokenizer, batch_input, input_kwargs) async with self.semaphore: return await asyncio.to_thread(self._get_scores, *input_args)
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/chat/vllm_engine.py
src/llamafactory/chat/vllm_engine.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import uuid from collections.abc import AsyncGenerator, AsyncIterator from typing import TYPE_CHECKING, Any, Optional, Union from packaging import version from typing_extensions import override from ..data import get_template_and_fix_tokenizer from ..extras import logging from ..extras.constants import AUDIO_PLACEHOLDER, IMAGE_PLACEHOLDER, VIDEO_PLACEHOLDER, EngineName from ..extras.misc import get_device_count from ..extras.packages import is_vllm_available from ..model import load_config, load_tokenizer from ..model.model_utils.quantization import QuantizationMethod from ..model.model_utils.visual import LlavaMultiModalProjectorForYiVLForVLLM from .base_engine import BaseEngine, Response if is_vllm_available(): from vllm import AsyncEngineArgs, AsyncLLMEngine, RequestOutput, SamplingParams from vllm.lora.request import LoRARequest if TYPE_CHECKING: from ..data.mm_plugin import AudioInput, ImageInput, VideoInput from ..hparams import DataArguments, FinetuningArguments, GeneratingArguments, ModelArguments logger = logging.get_logger(__name__) class VllmEngine(BaseEngine): def __init__( self, model_args: "ModelArguments", data_args: "DataArguments", finetuning_args: "FinetuningArguments", generating_args: "GeneratingArguments", ) -> None: self.name = EngineName.VLLM self.model_args = model_args config = load_config(model_args) # may download model from ms hub if getattr(config, "quantization_config", None): # gptq models should use float16 quantization_config: dict[str, Any] = getattr(config, "quantization_config", None) quant_method = quantization_config.get("quant_method", "") if quant_method == QuantizationMethod.GPTQ and model_args.infer_dtype == "auto": model_args.infer_dtype = "float16" self.can_generate = finetuning_args.stage == "sft" tokenizer_module = load_tokenizer(model_args) self.tokenizer = tokenizer_module["tokenizer"] self.processor = tokenizer_module["processor"] self.tokenizer.padding_side = "left" self.template = get_template_and_fix_tokenizer(self.tokenizer, data_args) self.template.mm_plugin.expand_mm_tokens = False # for vllm generate self.generating_args = generating_args.to_dict() engine_args = { "model": model_args.model_name_or_path, "trust_remote_code": model_args.trust_remote_code, "download_dir": model_args.cache_dir, "dtype": model_args.infer_dtype, "max_model_len": model_args.vllm_maxlen, "tensor_parallel_size": get_device_count() or 1, "gpu_memory_utilization": model_args.vllm_gpu_util, "disable_log_stats": True, "enforce_eager": model_args.vllm_enforce_eager, "enable_lora": model_args.adapter_name_or_path is not None, "max_lora_rank": model_args.vllm_max_lora_rank, } import vllm if version.parse(vllm.__version__) <= version.parse("0.10.0"): engine_args["disable_log_requests"] = True else: engine_args["enable_log_requests"] = False if self.template.mm_plugin.__class__.__name__ != "BasePlugin": engine_args["limit_mm_per_prompt"] = {"image": 4, "video": 2, "audio": 2} if isinstance(model_args.vllm_config, dict): engine_args.update(model_args.vllm_config) if getattr(config, "is_yi_vl_derived_model", None): import vllm.model_executor.models.llava logger.info_rank0("Detected Yi-VL model, applying projector patch.") vllm.model_executor.models.llava.LlavaMultiModalProjector = LlavaMultiModalProjectorForYiVLForVLLM self.model = AsyncLLMEngine.from_engine_args(AsyncEngineArgs(**engine_args)) if model_args.adapter_name_or_path is not None: self.lora_request = LoRARequest("default", 1, model_args.adapter_name_or_path[0]) else: self.lora_request = None async def _generate( self, messages: list[dict[str, str]], system: Optional[str] = None, tools: Optional[str] = None, images: Optional[list["ImageInput"]] = None, videos: Optional[list["VideoInput"]] = None, audios: Optional[list["AudioInput"]] = None, **input_kwargs, ) -> AsyncIterator["RequestOutput"]: request_id = f"chatcmpl-{uuid.uuid4().hex}" if images is not None and not any(IMAGE_PLACEHOLDER in message["content"] for message in messages): messages[0]["content"] = IMAGE_PLACEHOLDER * len(images) + messages[0]["content"] if videos is not None and not any(VIDEO_PLACEHOLDER in message["content"] for message in messages): messages[0]["content"] = VIDEO_PLACEHOLDER * len(videos) + messages[0]["content"] if audios is not None and not any(AUDIO_PLACEHOLDER in message["content"] for message in messages): messages[0]["content"] = AUDIO_PLACEHOLDER * len(audios) + messages[0]["content"] messages = self.template.mm_plugin.process_messages( messages, images or [], videos or [], audios or [], self.processor ) paired_messages = messages + [{"role": "assistant", "content": ""}] prompt_ids, _ = self.template.encode_oneturn(self.tokenizer, paired_messages, system, tools) prompt_length = len(prompt_ids) temperature: Optional[float] = input_kwargs.pop("temperature", None) top_p: Optional[float] = input_kwargs.pop("top_p", None) top_k: Optional[float] = input_kwargs.pop("top_k", None) num_return_sequences: int = input_kwargs.pop("num_return_sequences", 1) repetition_penalty: Optional[float] = input_kwargs.pop("repetition_penalty", None) length_penalty: Optional[float] = input_kwargs.pop("length_penalty", None) skip_special_tokens: Optional[bool] = input_kwargs.pop("skip_special_tokens", None) max_length: Optional[int] = input_kwargs.pop("max_length", None) max_new_tokens: Optional[int] = input_kwargs.pop("max_new_tokens", None) stop: Optional[Union[str, list[str]]] = input_kwargs.pop("stop", None) if length_penalty is not None: logger.warning_rank0("Length penalty is not supported by the vllm engine yet.") if "max_new_tokens" in self.generating_args: max_tokens = self.generating_args["max_new_tokens"] elif "max_length" in self.generating_args: if self.generating_args["max_length"] > prompt_length: max_tokens = self.generating_args["max_length"] - prompt_length else: max_tokens = 1 if max_length: max_tokens = max_length - prompt_length if max_length > prompt_length else 1 if max_new_tokens: max_tokens = max_new_tokens sampling_params = SamplingParams( n=num_return_sequences, repetition_penalty=( repetition_penalty if repetition_penalty is not None else self.generating_args["repetition_penalty"] ) or 1.0, # repetition_penalty must > 0 temperature=temperature if temperature is not None else self.generating_args["temperature"], top_p=(top_p if top_p is not None else self.generating_args["top_p"]) or 1.0, # top_p must > 0 top_k=(top_k if top_k is not None else self.generating_args["top_k"]) or -1, # top_k must > 0 stop=stop, stop_token_ids=self.template.get_stop_token_ids(self.tokenizer), max_tokens=max_tokens, skip_special_tokens=skip_special_tokens if skip_special_tokens is not None else self.generating_args["skip_special_tokens"], ) if images is not None: # add image features multi_modal_data = { "image": self.template.mm_plugin._regularize_images( images, image_max_pixels=self.model_args.image_max_pixels, image_min_pixels=self.model_args.image_min_pixels, )["images"] } elif videos is not None: multi_modal_data = { "video": self.template.mm_plugin._regularize_videos( videos, image_max_pixels=self.model_args.video_max_pixels, image_min_pixels=self.model_args.video_min_pixels, video_fps=self.model_args.video_fps, video_maxlen=self.model_args.video_maxlen, )["videos"] } elif audios is not None: audio_data = self.template.mm_plugin._regularize_audios( audios, sampling_rate=self.model_args.audio_sampling_rate, ) multi_modal_data = {"audio": zip(audio_data["audios"], audio_data["sampling_rates"])} else: multi_modal_data = None result_generator = self.model.generate( {"prompt_token_ids": prompt_ids, "multi_modal_data": multi_modal_data}, sampling_params=sampling_params, request_id=request_id, lora_request=self.lora_request, ) return result_generator @override async def chat( self, messages: list[dict[str, str]], system: Optional[str] = None, tools: Optional[str] = None, images: Optional[list["ImageInput"]] = None, videos: Optional[list["VideoInput"]] = None, audios: Optional[list["AudioInput"]] = None, **input_kwargs, ) -> list["Response"]: final_output = None generator = await self._generate(messages, system, tools, images, videos, audios, **input_kwargs) async for request_output in generator: final_output = request_output results = [] for output in final_output.outputs: results.append( Response( response_text=output.text, response_length=len(output.token_ids), prompt_length=len(final_output.prompt_token_ids), finish_reason=output.finish_reason, ) ) return results @override async def stream_chat( self, messages: list[dict[str, str]], system: Optional[str] = None, tools: Optional[str] = None, images: Optional[list["ImageInput"]] = None, videos: Optional[list["VideoInput"]] = None, audios: Optional[list["AudioInput"]] = None, **input_kwargs, ) -> AsyncGenerator[str, None]: generated_text = "" generator = await self._generate(messages, system, tools, images, videos, audios, **input_kwargs) async for result in generator: delta_text = result.outputs[0].text[len(generated_text) :] generated_text = result.outputs[0].text yield delta_text @override async def get_scores( self, batch_input: list[str], **input_kwargs, ) -> list[float]: raise NotImplementedError("vLLM engine does not support `get_scores`.")
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/chat/sglang_engine.py
src/llamafactory/chat/sglang_engine.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import asyncio import atexit import json from collections.abc import AsyncGenerator, AsyncIterator, Sequence from typing import TYPE_CHECKING, Any, Optional, Union import requests from typing_extensions import override from ..data import get_template_and_fix_tokenizer from ..extras import logging from ..extras.constants import AUDIO_PLACEHOLDER, IMAGE_PLACEHOLDER, VIDEO_PLACEHOLDER, EngineName from ..extras.misc import get_device_count, torch_gc from ..extras.packages import is_sglang_available from ..hparams import DataArguments, FinetuningArguments, GeneratingArguments, ModelArguments from ..model import load_config, load_tokenizer from ..model.model_utils.quantization import QuantizationMethod from .base_engine import BaseEngine, Response if is_sglang_available(): from sglang.utils import launch_server_cmd, terminate_process, wait_for_server # type: ignore if TYPE_CHECKING: from ..data.mm_plugin import AudioInput, ImageInput, VideoInput logger = logging.get_logger(__name__) class SGLangEngine(BaseEngine): """Inference engine for SGLang models. This class wraps the SGLang engine to provide a consistent interface for text generation that matches LLaMA Factory's requirements. It uses the SGLang HTTP server approach for better interaction and performance. The engine launches a server process and communicates with it via HTTP requests. For more details on the SGLang HTTP server approach, see: https://docs.sglang.ai/backend/send_request.html """ def __init__( self, model_args: "ModelArguments", data_args: "DataArguments", finetuning_args: "FinetuningArguments", generating_args: "GeneratingArguments", ) -> None: self.name = EngineName.SGLANG self.model_args = model_args config = load_config(model_args) # may download model from ms hub if getattr(config, "quantization_config", None): # gptq models should use float16 quantization_config: dict[str, Any] = getattr(config, "quantization_config", None) quant_method = quantization_config.get("quant_method", "") if quant_method == QuantizationMethod.GPTQ and model_args.infer_dtype == "auto": model_args.infer_dtype = "float16" self.can_generate = finetuning_args.stage == "sft" tokenizer_module = load_tokenizer(model_args) self.tokenizer = tokenizer_module["tokenizer"] self.processor = tokenizer_module["processor"] self.tokenizer.padding_side = "left" self.template = get_template_and_fix_tokenizer(self.tokenizer, data_args) self.template.mm_plugin.expand_mm_tokens = False # for sglang generate self.generating_args = generating_args.to_dict() if model_args.adapter_name_or_path is not None: self.lora_request = True else: self.lora_request = False launch_cmd = [ "python3 -m sglang.launch_server", f"--model-path {model_args.model_name_or_path}", f"--dtype {model_args.infer_dtype}", f"--context-length {model_args.sglang_maxlen}", f"--mem-fraction-static {model_args.sglang_mem_fraction}", f"--tp-size {model_args.sglang_tp_size if model_args.sglang_tp_size != -1 else get_device_count() or 1}", f"--download-dir {model_args.cache_dir}", "--log-level error", ] if self.lora_request: launch_cmd.extend( [ "--max-loras-per-batch 1", f"--lora-backend {model_args.sglang_lora_backend}", f"--lora-paths lora0={model_args.adapter_name_or_path[0]}", "--disable-radix-cache", ] ) launch_cmd = " ".join(launch_cmd) logger.info_rank0(f"Starting SGLang server with command: {launch_cmd}") try: torch_gc() self.server_process, port = launch_server_cmd(launch_cmd) self.base_url = f"http://localhost:{port}" atexit.register(self._cleanup_server) logger.info_rank0(f"Waiting for SGLang server to be ready at {self.base_url}") wait_for_server(self.base_url, timeout=300) logger.info_rank0(f"SGLang server initialized successfully at {self.base_url}") try: response = requests.get(f"{self.base_url}/get_model_info", timeout=5) if response.status_code == 200: model_info = response.json() logger.info(f"SGLang server model info: {model_info}") except Exception as e: logger.debug(f"Note: could not get model info: {str(e)}") except Exception as e: logger.error(f"Failed to start SGLang server: {str(e)}") self._cleanup_server() # make sure to clean up any started process raise RuntimeError(f"SGLang server initialization failed: {str(e)}.") def _cleanup_server(self): r"""Clean up the server process when the engine is destroyed.""" if hasattr(self, "server_process") and self.server_process: try: logger.info("Terminating SGLang server process") terminate_process(self.server_process) logger.info("SGLang server process terminated") except Exception as e: logger.warning(f"Error terminating SGLang server: {str(e)}") async def _generate( self, messages: list[dict[str, str]], system: Optional[str] = None, tools: Optional[str] = None, images: Optional[list["ImageInput"]] = None, videos: Optional[list["VideoInput"]] = None, audios: Optional[list["AudioInput"]] = None, **input_kwargs, ) -> AsyncIterator[dict[str, Any]]: if images is not None and not any(IMAGE_PLACEHOLDER in message["content"] for message in messages): messages[0]["content"] = IMAGE_PLACEHOLDER * len(images) + messages[0]["content"] if videos is not None and not any(VIDEO_PLACEHOLDER in message["content"] for message in messages): messages[0]["content"] = VIDEO_PLACEHOLDER * len(videos) + messages[0]["content"] if audios is not None and not any(AUDIO_PLACEHOLDER in message["content"] for message in messages): messages[0]["content"] = AUDIO_PLACEHOLDER * len(audios) + messages[0]["content"] messages = self.template.mm_plugin.process_messages( messages, images or [], videos or [], audios or [], self.processor ) paired_messages = messages + [{"role": "assistant", "content": ""}] prompt_ids, _ = self.template.encode_oneturn(self.tokenizer, paired_messages, system, tools) prompt_length = len(prompt_ids) temperature: Optional[float] = input_kwargs.pop("temperature", None) top_p: Optional[float] = input_kwargs.pop("top_p", None) top_k: Optional[float] = input_kwargs.pop("top_k", None) num_return_sequences: int = input_kwargs.pop("num_return_sequences", 1) repetition_penalty: Optional[float] = input_kwargs.pop("repetition_penalty", None) skip_special_tokens: Optional[bool] = input_kwargs.pop("skip_special_tokens", None) max_length: Optional[int] = input_kwargs.pop("max_length", None) max_new_tokens: Optional[int] = input_kwargs.pop("max_new_tokens", None) stop: Optional[Union[str, list[str]]] = input_kwargs.pop("stop", None) if num_return_sequences != 1: raise NotImplementedError("SGLang only supports n=1.") if "max_new_tokens" in self.generating_args: max_tokens = self.generating_args["max_new_tokens"] elif "max_length" in self.generating_args: if self.generating_args["max_length"] > prompt_length: max_tokens = self.generating_args["max_length"] - prompt_length else: max_tokens = 1 if max_length: max_tokens = max_length - prompt_length if max_length > prompt_length else 1 if max_new_tokens: max_tokens = max_new_tokens sampling_params = { "temperature": temperature if temperature is not None else self.generating_args["temperature"], "top_p": (top_p if top_p is not None else self.generating_args["top_p"]) or 1.0, # top_p must > 0 "top_k": (top_k if top_k is not None else self.generating_args["top_k"]) or -1, # top_k must > 0 "stop": stop, "stop_token_ids": self.template.get_stop_token_ids(self.tokenizer), "max_new_tokens": max_tokens, "repetition_penalty": ( repetition_penalty if repetition_penalty is not None else self.generating_args["repetition_penalty"] ) or 1.0, # repetition_penalty must > 0 "skip_special_tokens": skip_special_tokens if skip_special_tokens is not None else self.generating_args["skip_special_tokens"], } def stream_request(): json_data = { "input_ids": prompt_ids, "sampling_params": sampling_params, "stream": True, } if self.lora_request: json_data["lora_request"] = ["lora0"] response = requests.post(f"{self.base_url}/generate", json=json_data, stream=True) if response.status_code != 200: raise RuntimeError(f"SGLang server error: {response.status_code}, {response.text}") for chunk in response.iter_lines(decode_unicode=False): chunk = str(chunk.decode("utf-8")) if chunk == "data: [DONE]": break if chunk and chunk.startswith("data:"): yield json.loads(chunk[5:].strip("\n")) return await asyncio.to_thread(stream_request) @override async def chat( self, messages: Sequence[dict[str, str]], system: Optional[str] = None, tools: Optional[str] = None, images: Optional[Sequence["ImageInput"]] = None, videos: Optional[Sequence["VideoInput"]] = None, audios: Optional[Sequence["AudioInput"]] = None, **input_kwargs, ) -> list["Response"]: final_output = None generator = await self._generate(messages, system, tools, images, videos, audios, **input_kwargs) for request_output in generator: final_output = request_output results = [ Response( response_text=final_output["text"], response_length=final_output["meta_info"]["completion_tokens"], prompt_length=final_output["meta_info"]["prompt_tokens"], finish_reason="stop" if final_output["meta_info"]["finish_reason"] == "stop" else "length", ) ] return results @override async def stream_chat( self, messages: list[dict[str, str]], system: Optional[str] = None, tools: Optional[str] = None, images: Optional[list["ImageInput"]] = None, videos: Optional[list["VideoInput"]] = None, audios: Optional[list["AudioInput"]] = None, **input_kwargs, ) -> AsyncGenerator[str, None]: generated_text = "" generator = await self._generate(messages, system, tools, images, videos, audios, **input_kwargs) for result in generator: delta_text = result["text"][len(generated_text) :] generated_text = result["text"] yield delta_text @override async def get_scores( self, batch_input: list[str], **input_kwargs, ) -> list[float]: raise NotImplementedError("SGLang engine does not support `get_scores`.") def __del__(self): r"""Ensure server is cleaned up when object is deleted.""" self._cleanup_server() try: atexit.unregister(self._cleanup_server) except Exception: pass
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/chat/chat_model.py
src/llamafactory/chat/chat_model.py
# Copyright 2025 THUDM and the LlamaFactory team. # # This code is inspired by the THUDM's ChatGLM implementation. # https://github.com/THUDM/ChatGLM-6B/blob/main/cli_demo.py # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import asyncio import os from collections.abc import AsyncGenerator, Generator from threading import Thread from typing import TYPE_CHECKING, Any, Optional from ..extras.constants import EngineName from ..extras.misc import torch_gc from ..hparams import get_infer_args if TYPE_CHECKING: from ..data.mm_plugin import AudioInput, ImageInput, VideoInput from .base_engine import BaseEngine, Response def _start_background_loop(loop: "asyncio.AbstractEventLoop") -> None: asyncio.set_event_loop(loop) loop.run_forever() class ChatModel: r"""General class for chat models. Backed by huggingface or vllm engines. Supports both sync and async methods. Sync methods: chat(), stream_chat() and get_scores(). Async methods: achat(), astream_chat() and aget_scores(). """ def __init__(self, args: Optional[dict[str, Any]] = None) -> None: model_args, data_args, finetuning_args, generating_args = get_infer_args(args) if model_args.infer_backend == EngineName.HF: from .hf_engine import HuggingfaceEngine self.engine: BaseEngine = HuggingfaceEngine(model_args, data_args, finetuning_args, generating_args) elif model_args.infer_backend == EngineName.VLLM: try: from .vllm_engine import VllmEngine self.engine: BaseEngine = VllmEngine(model_args, data_args, finetuning_args, generating_args) except ImportError as e: raise ImportError( "vLLM not install, you may need to run `pip install vllm`\n" "or try to use HuggingFace backend: --infer_backend huggingface" ) from e elif model_args.infer_backend == EngineName.SGLANG: try: from .sglang_engine import SGLangEngine self.engine: BaseEngine = SGLangEngine(model_args, data_args, finetuning_args, generating_args) except ImportError as e: raise ImportError( "SGLang not install, you may need to run `pip install sglang[all]`\n" "or try to use HuggingFace backend: --infer_backend huggingface" ) from e elif model_args.infer_backend == EngineName.KT: try: from .kt_engine import KTransformersEngine self.engine: BaseEngine = KTransformersEngine(model_args, data_args, finetuning_args, generating_args) except ImportError as e: raise ImportError( "KTransformers not install, you may need to run `pip install ktransformers`\n" "or try to use HuggingFace backend: --infer_backend huggingface" ) from e else: raise NotImplementedError(f"Unknown backend: {model_args.infer_backend}") self._loop = asyncio.new_event_loop() self._thread = Thread(target=_start_background_loop, args=(self._loop,), daemon=True) self._thread.start() def chat( self, messages: list[dict[str, str]], system: Optional[str] = None, tools: Optional[str] = None, images: Optional[list["ImageInput"]] = None, videos: Optional[list["VideoInput"]] = None, audios: Optional[list["AudioInput"]] = None, **input_kwargs, ) -> list["Response"]: r"""Get a list of responses of the chat model.""" task = asyncio.run_coroutine_threadsafe( self.achat(messages, system, tools, images, videos, audios, **input_kwargs), self._loop ) return task.result() async def achat( self, messages: list[dict[str, str]], system: Optional[str] = None, tools: Optional[str] = None, images: Optional[list["ImageInput"]] = None, videos: Optional[list["VideoInput"]] = None, audios: Optional[list["AudioInput"]] = None, **input_kwargs, ) -> list["Response"]: r"""Asynchronously get a list of responses of the chat model.""" return await self.engine.chat(messages, system, tools, images, videos, audios, **input_kwargs) def stream_chat( self, messages: list[dict[str, str]], system: Optional[str] = None, tools: Optional[str] = None, images: Optional[list["ImageInput"]] = None, videos: Optional[list["VideoInput"]] = None, audios: Optional[list["AudioInput"]] = None, **input_kwargs, ) -> Generator[str, None, None]: r"""Get the response token-by-token of the chat model.""" generator = self.astream_chat(messages, system, tools, images, videos, audios, **input_kwargs) while True: try: task = asyncio.run_coroutine_threadsafe(generator.__anext__(), self._loop) yield task.result() except StopAsyncIteration: break async def astream_chat( self, messages: list[dict[str, str]], system: Optional[str] = None, tools: Optional[str] = None, images: Optional[list["ImageInput"]] = None, videos: Optional[list["VideoInput"]] = None, audios: Optional[list["AudioInput"]] = None, **input_kwargs, ) -> AsyncGenerator[str, None]: r"""Asynchronously get the response token-by-token of the chat model.""" async for new_token in self.engine.stream_chat( messages, system, tools, images, videos, audios, **input_kwargs ): yield new_token def get_scores( self, batch_input: list[str], **input_kwargs, ) -> list[float]: r"""Get a list of scores of the reward model.""" task = asyncio.run_coroutine_threadsafe(self.aget_scores(batch_input, **input_kwargs), self._loop) return task.result() async def aget_scores( self, batch_input: list[str], **input_kwargs, ) -> list[float]: r"""Asynchronously get a list of scores of the reward model.""" return await self.engine.get_scores(batch_input, **input_kwargs) def run_chat() -> None: if os.name != "nt": try: import readline # noqa: F401 except ImportError: print("Install `readline` for a better experience.") chat_model = ChatModel() messages = [] print("Welcome to the CLI application, use `clear` to remove the history, use `exit` to exit the application.") while True: try: query = input("\nUser: ") except UnicodeDecodeError: print("Detected decoding error at the inputs, please set the terminal encoding to utf-8.") continue except Exception: raise if query.strip() == "exit": break if query.strip() == "clear": messages = [] torch_gc() print("History has been removed.") continue messages.append({"role": "user", "content": query}) print("Assistant: ", end="", flush=True) response = "" for new_text in chat_model.stream_chat(messages): print(new_text, end="", flush=True) response += new_text print() messages.append({"role": "assistant", "content": response})
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/chat/__init__.py
src/llamafactory/chat/__init__.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from .base_engine import BaseEngine from .chat_model import ChatModel __all__ = ["BaseEngine", "ChatModel"]
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/chat/base_engine.py
src/llamafactory/chat/base_engine.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from abc import ABC, abstractmethod from collections.abc import AsyncGenerator from dataclasses import dataclass from typing import TYPE_CHECKING, Any, Literal, Optional, Union if TYPE_CHECKING: from transformers import PreTrainedModel, PreTrainedTokenizer from vllm import AsyncLLMEngine from ..data import Template from ..data.mm_plugin import AudioInput, ImageInput, VideoInput from ..extras.constants import EngineName from ..hparams import DataArguments, FinetuningArguments, GeneratingArguments, ModelArguments @dataclass class Response: response_text: str response_length: int prompt_length: int finish_reason: Literal["stop", "length"] class BaseEngine(ABC): r"""Base class for inference engine of chat models. Must implements async methods: chat(), stream_chat() and get_scores(). """ name: "EngineName" model: Union["PreTrainedModel", "AsyncLLMEngine"] tokenizer: "PreTrainedTokenizer" can_generate: bool template: "Template" generating_args: dict[str, Any] @abstractmethod def __init__( self, model_args: "ModelArguments", data_args: "DataArguments", finetuning_args: "FinetuningArguments", generating_args: "GeneratingArguments", ) -> None: r"""Initialize an inference engine.""" ... @abstractmethod async def chat( self, messages: list[dict[str, str]], system: Optional[str] = None, tools: Optional[str] = None, images: Optional[list["ImageInput"]] = None, videos: Optional[list["VideoInput"]] = None, audios: Optional[list["AudioInput"]] = None, **input_kwargs, ) -> list["Response"]: r"""Get a list of responses of the chat model.""" ... @abstractmethod async def stream_chat( self, messages: list[dict[str, str]], system: Optional[str] = None, tools: Optional[str] = None, images: Optional[list["ImageInput"]] = None, videos: Optional[list["VideoInput"]] = None, audios: Optional[list["AudioInput"]] = None, **input_kwargs, ) -> AsyncGenerator[str, None]: r"""Get the response token-by-token of the chat model.""" ... @abstractmethod async def get_scores( self, batch_input: list[str], **input_kwargs, ) -> list[float]: r"""Get a list of scores of the reward model.""" ...
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/train/callbacks.py
src/llamafactory/train/callbacks.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import json import os import signal import sys import time from concurrent.futures import ThreadPoolExecutor from datetime import timedelta from typing import TYPE_CHECKING, Any, Optional import torch import transformers from peft import PeftModel from transformers import PreTrainedModel, ProcessorMixin, TrainerCallback from transformers.trainer_utils import PREFIX_CHECKPOINT_DIR, has_length from transformers.utils import SAFE_WEIGHTS_NAME, WEIGHTS_NAME from typing_extensions import override from ..extras import logging from ..extras.constants import TRAINER_LOG, V_HEAD_SAFE_WEIGHTS_NAME, V_HEAD_WEIGHTS_NAME from ..extras.misc import get_peak_memory, is_env_enabled, use_ray from ..extras.packages import is_safetensors_available if is_safetensors_available(): from safetensors import safe_open from safetensors.torch import save_file if TYPE_CHECKING: from transformers import TrainerControl, TrainerState, TrainingArguments from trl import AutoModelForCausalLMWithValueHead from ..hparams import DataArguments, FinetuningArguments, GeneratingArguments, ModelArguments logger = logging.get_logger(__name__) def fix_valuehead_checkpoint( model: "AutoModelForCausalLMWithValueHead", output_dir: str, safe_serialization: bool ) -> None: r"""Fix the valuehead checkpoint files. The model is already unwrapped. There are three cases: 1. full tuning without ds_zero3: state_dict = {"model.layers.*": ..., "v_head.summary.*": ...} 2. lora tuning without ds_zero3: state_dict = {"v_head.summary.*": ...} 3. under deepspeed zero3: state_dict = {"pretrained_model.model.layers.*": ..., "v_head.summary.*": ...} We assume `stage3_gather_16bit_weights_on_model_save=true`. """ if not isinstance(model.pretrained_model, (PreTrainedModel, PeftModel)): return if safe_serialization: path_to_checkpoint = os.path.join(output_dir, SAFE_WEIGHTS_NAME) with safe_open(path_to_checkpoint, framework="pt", device="cpu") as f: state_dict: dict[str, torch.Tensor] = {key: f.get_tensor(key).clone() for key in f.keys()} else: path_to_checkpoint = os.path.join(output_dir, WEIGHTS_NAME) state_dict: dict[str, torch.Tensor] = torch.load(path_to_checkpoint, map_location="cpu", weights_only=True) os.remove(path_to_checkpoint) decoder_state_dict, v_head_state_dict = {}, {} for name, param in state_dict.items(): if name.startswith("v_head."): v_head_state_dict[name] = param else: decoder_state_dict[name.replace("pretrained_model.", "", 1)] = param model.pretrained_model.save_pretrained( output_dir, state_dict=decoder_state_dict or None, safe_serialization=safe_serialization ) if safe_serialization: save_file(v_head_state_dict, os.path.join(output_dir, V_HEAD_SAFE_WEIGHTS_NAME), metadata={"format": "pt"}) else: torch.save(v_head_state_dict, os.path.join(output_dir, V_HEAD_WEIGHTS_NAME)) logger.info_rank0(f"Value head model saved at: {output_dir}") class FixValueHeadModelCallback(TrainerCallback): r"""A callback for fixing the checkpoint for valuehead models.""" @override def on_save(self, args: "TrainingArguments", state: "TrainerState", control: "TrainerControl", **kwargs): if args.should_save: output_dir = os.path.join(args.output_dir, f"{PREFIX_CHECKPOINT_DIR}-{state.global_step}") fix_valuehead_checkpoint( model=kwargs.pop("model"), output_dir=output_dir, safe_serialization=args.save_safetensors ) class SaveProcessorCallback(TrainerCallback): r"""A callback for saving the processor.""" def __init__(self, processor: "ProcessorMixin") -> None: self.processor = processor @override def on_save(self, args: "TrainingArguments", state: "TrainerState", control: "TrainerControl", **kwargs): if args.should_save: output_dir = os.path.join(args.output_dir, f"{PREFIX_CHECKPOINT_DIR}-{state.global_step}") self.processor.save_pretrained(output_dir) @override def on_train_end(self, args: "TrainingArguments", state: "TrainerState", control: "TrainerControl", **kwargs): if args.should_save: self.processor.save_pretrained(args.output_dir) class PissaConvertCallback(TrainerCallback): r"""A callback for converting the PiSSA adapter to a normal one.""" @override def on_train_begin(self, args: "TrainingArguments", state: "TrainerState", control: "TrainerControl", **kwargs): if args.should_save: model = kwargs.pop("model") pissa_init_dir = os.path.join(args.output_dir, "pissa_init") logger.info_rank0(f"Initial PiSSA adapter will be saved at: {pissa_init_dir}.") if isinstance(model, PeftModel): init_lora_weights = getattr(model.peft_config["default"], "init_lora_weights") setattr(model.peft_config["default"], "init_lora_weights", True) model.save_pretrained(pissa_init_dir, safe_serialization=args.save_safetensors) setattr(model.peft_config["default"], "init_lora_weights", init_lora_weights) @override def on_train_end(self, args: "TrainingArguments", state: "TrainerState", control: "TrainerControl", **kwargs): if args.should_save: model = kwargs.pop("model") pissa_init_dir = os.path.join(args.output_dir, "pissa_init") pissa_backup_dir = os.path.join(args.output_dir, "pissa_backup") pissa_convert_dir = os.path.join(args.output_dir, "pissa_converted") logger.info_rank0(f"Converted PiSSA adapter will be saved at: {pissa_convert_dir}.") # 1. save a pissa backup with init_lora_weights: True # 2. save a converted lora with init_lora_weights: pissa # 3. load the pissa backup with init_lora_weights: True # 4. delete the initial adapter and change init_lora_weights to pissa if isinstance(model, PeftModel): init_lora_weights = getattr(model.peft_config["default"], "init_lora_weights") setattr(model.peft_config["default"], "init_lora_weights", True) model.save_pretrained(pissa_backup_dir, safe_serialization=args.save_safetensors) setattr(model.peft_config["default"], "init_lora_weights", init_lora_weights) model.save_pretrained( pissa_convert_dir, safe_serialization=args.save_safetensors, path_initial_model_for_weight_conversion=pissa_init_dir, ) model.load_adapter(pissa_backup_dir, "default", is_trainable=True) model.set_adapter("default") setattr(model.peft_config["default"], "init_lora_weights", init_lora_weights) class LogCallback(TrainerCallback): r"""A callback for logging training and evaluation status.""" def __init__(self) -> None: # Progress self.start_time = 0 self.cur_steps = 0 self.max_steps = 0 self.elapsed_time = "" self.remaining_time = "" self.thread_pool: Optional[ThreadPoolExecutor] = None # Status self.aborted = False self.do_train = False # Web UI self.webui_mode = is_env_enabled("LLAMABOARD_ENABLED") if self.webui_mode and not use_ray(): signal.signal(signal.SIGABRT, self._set_abort) self.logger_handler = logging.LoggerHandler(os.getenv("LLAMABOARD_WORKDIR")) logging.add_handler(self.logger_handler) transformers.logging.add_handler(self.logger_handler) def _set_abort(self, signum, frame) -> None: self.aborted = True def _reset(self, max_steps: int = 0) -> None: self.start_time = time.time() self.cur_steps = 0 self.max_steps = max_steps self.elapsed_time = "" self.remaining_time = "" def _timing(self, cur_steps: int) -> None: cur_time = time.time() elapsed_time = cur_time - self.start_time avg_time_per_step = elapsed_time / cur_steps if cur_steps != 0 else 0 remaining_time = (self.max_steps - cur_steps) * avg_time_per_step self.cur_steps = cur_steps self.elapsed_time = str(timedelta(seconds=int(elapsed_time))) self.remaining_time = str(timedelta(seconds=int(remaining_time))) def _write_log(self, output_dir: str, logs: dict[str, Any]) -> None: with open(os.path.join(output_dir, TRAINER_LOG), "a", encoding="utf-8") as f: f.write(json.dumps(logs) + "\n") def _create_thread_pool(self, output_dir: str) -> None: os.makedirs(output_dir, exist_ok=True) self.thread_pool = ThreadPoolExecutor(max_workers=1) def _close_thread_pool(self) -> None: if self.thread_pool is not None: self.thread_pool.shutdown(wait=True) self.thread_pool = None @override def on_init_end(self, args: "TrainingArguments", state: "TrainerState", control: "TrainerControl", **kwargs): if ( args.should_save and os.path.exists(os.path.join(args.output_dir, TRAINER_LOG)) and args.overwrite_output_dir ): logger.warning_rank0_once("Previous trainer log in this folder will be deleted.") os.remove(os.path.join(args.output_dir, TRAINER_LOG)) @override def on_train_begin(self, args: "TrainingArguments", state: "TrainerState", control: "TrainerControl", **kwargs): if args.should_save: self.do_train = True self._reset(max_steps=state.max_steps) self._create_thread_pool(output_dir=args.output_dir) @override def on_train_end(self, args: "TrainingArguments", state: "TrainerState", control: "TrainerControl", **kwargs): self._close_thread_pool() @override def on_substep_end(self, args: "TrainingArguments", state: "TrainerState", control: "TrainerControl", **kwargs): if self.aborted: control.should_epoch_stop = True control.should_training_stop = True @override def on_step_end(self, args: "TrainingArguments", state: "TrainerState", control: "TrainerControl", **kwargs): if self.aborted: control.should_epoch_stop = True control.should_training_stop = True @override def on_evaluate(self, args: "TrainingArguments", state: "TrainerState", control: "TrainerControl", **kwargs): if not self.do_train: self._close_thread_pool() @override def on_predict(self, args: "TrainingArguments", state: "TrainerState", control: "TrainerControl", **kwargs): if not self.do_train: self._close_thread_pool() @override def on_log(self, args: "TrainingArguments", state: "TrainerState", control: "TrainerControl", **kwargs): if not args.should_save: return self._timing(cur_steps=state.global_step) logs = dict( current_steps=self.cur_steps, total_steps=self.max_steps, loss=state.log_history[-1].get("loss"), eval_loss=state.log_history[-1].get("eval_loss"), predict_loss=state.log_history[-1].get("predict_loss"), reward=state.log_history[-1].get("reward"), accuracy=state.log_history[-1].get("rewards/accuracies"), lr=state.log_history[-1].get("learning_rate"), epoch=state.log_history[-1].get("epoch"), percentage=round(self.cur_steps / self.max_steps * 100, 2) if self.max_steps != 0 else 100, elapsed_time=self.elapsed_time, remaining_time=self.remaining_time, ) if state.num_input_tokens_seen: logs["throughput"] = round(state.num_input_tokens_seen / (time.time() - self.start_time), 2) logs["total_tokens"] = state.num_input_tokens_seen if is_env_enabled("RECORD_VRAM"): vram_allocated, vram_reserved = get_peak_memory() logs["vram_allocated"] = round(vram_allocated / (1024**3), 2) logs["vram_reserved"] = round(vram_reserved / (1024**3), 2) logs = {k: v for k, v in logs.items() if v is not None} if self.webui_mode and all(key in logs for key in ("loss", "lr", "epoch")): log_str = f"'loss': {logs['loss']:.4f}, 'learning_rate': {logs['lr']:2.4e}, 'epoch': {logs['epoch']:.2f}" for extra_key in ("reward", "accuracy", "throughput"): if logs.get(extra_key): log_str += f", '{extra_key}': {logs[extra_key]:.2f}" logger.info_rank0("{" + log_str + "}") if self.thread_pool is not None: self.thread_pool.submit(self._write_log, args.output_dir, logs) @override def on_prediction_step( self, args: "TrainingArguments", state: "TrainerState", control: "TrainerControl", **kwargs ): if self.do_train: return if self.aborted: sys.exit(0) if not args.should_save: return eval_dataloader = kwargs.pop("eval_dataloader", None) if has_length(eval_dataloader): if self.max_steps == 0: self._reset(max_steps=len(eval_dataloader)) self._create_thread_pool(output_dir=args.output_dir) self._timing(cur_steps=self.cur_steps + 1) if self.cur_steps % 5 == 0 and self.thread_pool is not None: logs = dict( current_steps=self.cur_steps, total_steps=self.max_steps, percentage=round(self.cur_steps / self.max_steps * 100, 2) if self.max_steps != 0 else 100, elapsed_time=self.elapsed_time, remaining_time=self.remaining_time, ) self.thread_pool.submit(self._write_log, args.output_dir, logs) class ReporterCallback(TrainerCallback): r"""A callback for reporting training status to external logger.""" def __init__( self, model_args: "ModelArguments", data_args: "DataArguments", finetuning_args: "FinetuningArguments", generating_args: "GeneratingArguments", ) -> None: self.model_args = model_args self.data_args = data_args self.finetuning_args = finetuning_args self.generating_args = generating_args os.environ["WANDB_PROJECT"] = os.getenv("WANDB_PROJECT", "llamafactory") @override def on_train_begin(self, args: "TrainingArguments", state: "TrainerState", control: "TrainerControl", **kwargs): if not state.is_world_process_zero: return if "wandb" in args.report_to: import wandb wandb.config.update( { "model_args": self.model_args.to_dict(), "data_args": self.data_args.to_dict(), "finetuning_args": self.finetuning_args.to_dict(), "generating_args": self.generating_args.to_dict(), } ) if self.finetuning_args.use_swanlab: import swanlab # type: ignore swanlab.config.update( { "model_args": self.model_args.to_dict(), "data_args": self.data_args.to_dict(), "finetuning_args": self.finetuning_args.to_dict(), "generating_args": self.generating_args.to_dict(), } )
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/train/trainer_utils.py
src/llamafactory/train/trainer_utils.py
# Copyright 2025 HuggingFace Inc. and the LlamaFactory team. # # This code is inspired by the original GaLore's implementation: https://github.com/jiaweizzhao/GaLore # and the original LoRA+'s implementation: https://github.com/nikhil-ghosh-berkeley/loraplus # and the original BAdam's implementation: https://github.com/Ledzy/BAdam # and the HuggingFace's TRL library: https://github.com/huggingface/trl # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import json import os from collections.abc import Callable, Mapping from pathlib import Path from typing import TYPE_CHECKING, Any, Optional, Union import torch from transformers import Trainer from transformers.integrations import is_deepspeed_zero3_enabled from transformers.modeling_utils import is_fsdp_enabled from transformers.optimization import get_scheduler from transformers.pytorch_utils import ALL_LAYERNORM_LAYERS from transformers.trainer_pt_utils import get_parameter_names from typing_extensions import override from ..extras import logging from ..extras.constants import IGNORE_INDEX, SWANLAB_CONFIG from ..extras.packages import is_apollo_available, is_galore_available, is_ray_available from ..hparams import FinetuningArguments, ModelArguments from ..model import find_all_linear_modules, load_model, load_tokenizer, load_valuehead_params if is_galore_available(): from galore_torch import GaLoreAdafactor, GaLoreAdamW, GaLoreAdamW8bit # type: ignore if is_apollo_available(): from apollo_torch import APOLLOAdamW # type: ignore if is_ray_available(): import ray from ray.train import RunConfig, ScalingConfig from ray.train.torch import TorchTrainer if TYPE_CHECKING: from transformers import PreTrainedModel, TrainerCallback, TrainerState from trl import AutoModelForCausalLMWithValueHead from ..hparams import DataArguments, RayArguments, TrainingArguments logger = logging.get_logger(__name__) class DummyOptimizer(torch.optim.Optimizer): r"""A dummy optimizer used for the GaLore or APOLLO algorithm.""" def __init__( self, lr: float = 1e-3, optimizer_dict: Optional[dict["torch.nn.Parameter", "torch.optim.Optimizer"]] = None ) -> None: dummy_tensor = torch.randn(1, 1) self.optimizer_dict = optimizer_dict super().__init__([dummy_tensor], {"lr": lr}) @override def zero_grad(self, set_to_none: bool = True) -> None: pass @override def step(self, closure: Optional[Callable[[], float]] = None) -> Optional[float]: pass def create_modelcard_and_push( trainer: "Trainer", model_args: "ModelArguments", data_args: "DataArguments", training_args: "TrainingArguments", finetuning_args: "FinetuningArguments", ) -> None: kwargs = { "tasks": "text-generation", "finetuned_from": model_args.model_name_or_path, "tags": ["llama-factory", finetuning_args.finetuning_type], } if data_args.dataset is not None: kwargs["dataset"] = data_args.dataset if model_args.use_unsloth: kwargs["tags"] = kwargs["tags"] + ["unsloth"] if model_args.use_kt: kwargs["tags"] = kwargs["tags"] + ["ktransformers"] if not training_args.do_train: pass elif training_args.push_to_hub: trainer.push_to_hub(**kwargs) else: Trainer.create_model_card(trainer, license="other", **kwargs) # prevent from connecting to hub def create_ref_model( model_args: "ModelArguments", finetuning_args: "FinetuningArguments", add_valuehead: bool = False ) -> Optional[Union["PreTrainedModel", "AutoModelForCausalLMWithValueHead"]]: r"""Create reference model for PPO/DPO training. Evaluation mode is not supported. The valuehead parameter is randomly initialized since it is useless for PPO training. """ if finetuning_args.ref_model is not None: ref_model_args = ModelArguments.copyfrom( model_args, model_name_or_path=finetuning_args.ref_model, adapter_name_or_path=finetuning_args.ref_model_adapters, quantization_bit=finetuning_args.ref_model_quantization_bit, ) ref_finetuning_args = FinetuningArguments() tokenizer = load_tokenizer(ref_model_args)["tokenizer"] ref_model = load_model( tokenizer, ref_model_args, ref_finetuning_args, is_trainable=False, add_valuehead=add_valuehead ) logger.info_rank0(f"Created reference model from {finetuning_args.ref_model}") else: if finetuning_args.finetuning_type == "lora": ref_model = None else: ref_model_args = ModelArguments.copyfrom(model_args) ref_finetuning_args = FinetuningArguments() tokenizer = load_tokenizer(ref_model_args)["tokenizer"] ref_model = load_model( tokenizer, ref_model_args, ref_finetuning_args, is_trainable=False, add_valuehead=add_valuehead ) logger.info_rank0("Created reference model from the model itself.") return ref_model def create_reward_model( model: "AutoModelForCausalLMWithValueHead", model_args: "ModelArguments", finetuning_args: "FinetuningArguments" ) -> Optional["AutoModelForCausalLMWithValueHead"]: r"""Create reward model for PPO training.""" if finetuning_args.reward_model_type == "api": assert finetuning_args.reward_model.startswith("http"), "Please provide full url." logger.info_rank0(f"Use reward server {finetuning_args.reward_model}") return finetuning_args.reward_model elif finetuning_args.reward_model_type == "lora": model.pretrained_model.load_adapter(finetuning_args.reward_model, "reward") for name, param in model.named_parameters(): # https://github.com/huggingface/peft/issues/1090 if "default" in name: param.data = param.data.to(torch.float32) # trainable params should in fp32 vhead_params = load_valuehead_params(finetuning_args.reward_model, model_args) assert vhead_params is not None, "Reward model is not correctly loaded." model.register_buffer("reward_head_weight", vhead_params["v_head.summary.weight"], persistent=False) model.register_buffer("reward_head_bias", vhead_params["v_head.summary.bias"], persistent=False) model.register_buffer( "default_head_weight", torch.zeros_like(vhead_params["v_head.summary.weight"]), persistent=False ) model.register_buffer( "default_head_bias", torch.zeros_like(vhead_params["v_head.summary.bias"]), persistent=False ) logger.info_rank0(f"Loaded adapter weights of reward model from {finetuning_args.reward_model}") return None else: reward_model_args = ModelArguments.copyfrom( model_args, model_name_or_path=finetuning_args.reward_model, adapter_name_or_path=finetuning_args.reward_model_adapters, quantization_bit=finetuning_args.reward_model_quantization_bit, ) reward_finetuning_args = FinetuningArguments() tokenizer = load_tokenizer(reward_model_args)["tokenizer"] reward_model = load_model( tokenizer, reward_model_args, reward_finetuning_args, is_trainable=False, add_valuehead=True ) logger.info_rank0(f"Loaded full weights of reward model from {finetuning_args.reward_model}") logger.warning_rank0("Please ensure the ppo model and reward model share SAME tokenizer and vocabulary.") return reward_model def _get_decay_parameter_names(model: "PreTrainedModel") -> list[str]: r"""Return a list of names of parameters with weight decay. (weights in non-layernorm layers).""" decay_parameters = get_parameter_names(model, ALL_LAYERNORM_LAYERS) decay_parameters = [name for name in decay_parameters if "bias" not in name] return decay_parameters def _create_galore_optimizer( model: "PreTrainedModel", training_args: "TrainingArguments", finetuning_args: "FinetuningArguments", ) -> "torch.optim.Optimizer": if len(finetuning_args.galore_target) == 1 and finetuning_args.galore_target[0] == "all": galore_targets = find_all_linear_modules(model, finetuning_args.freeze_vision_tower) else: galore_targets = finetuning_args.galore_target galore_params: list[torch.nn.Parameter] = [] for name, module in model.named_modules(): if isinstance(module, torch.nn.Linear) and any(target in name for target in galore_targets): for param in module.parameters(): if param.requires_grad and len(param.shape) > 1: galore_params.append(param) galore_kwargs = { "rank": finetuning_args.galore_rank, "update_proj_gap": finetuning_args.galore_update_interval, "scale": finetuning_args.galore_scale, "proj_type": finetuning_args.galore_proj_type, } id_galore_params = {id(param) for param in galore_params} decay_params, nodecay_params = [], [] # they are non-galore parameters trainable_params: list[torch.nn.Parameter] = [] # galore_params + decay_params + nodecay_params decay_param_names = _get_decay_parameter_names(model) for name, param in model.named_parameters(): if param.requires_grad: trainable_params.append(param) if id(param) not in id_galore_params: if name in decay_param_names: decay_params.append(param) else: nodecay_params.append(param) _, optim_kwargs = Trainer.get_optimizer_cls_and_kwargs(training_args) if training_args.optim == "adamw_torch": optim_class = GaLoreAdamW elif training_args.optim in ["adamw_bnb_8bit", "adamw_8bit", "paged_adamw_8bit"]: optim_class = GaLoreAdamW8bit elif training_args.optim == "adafactor": optim_class = GaLoreAdafactor else: raise NotImplementedError(f"Unknown optim: {training_args.optim}.") if finetuning_args.galore_layerwise: logger.warning_rank0("The displayed gradient norm will be all zeros in layerwise GaLore.") if training_args.gradient_accumulation_steps != 1: raise ValueError("Per-layer GaLore does not support gradient accumulation.") optimizer_dict: dict[torch.Tensor, torch.optim.Optimizer] = {} for param in nodecay_params: param_groups = [dict(params=[param], weight_decay=0.0)] optimizer_dict[param] = optim_class(param_groups, **optim_kwargs) for param in decay_params: param_groups = [dict(params=[param], weight_decay=training_args.weight_decay)] optimizer_dict[param] = optim_class(param_groups, **optim_kwargs) for param in galore_params: # galore params have weight decay param_groups = [dict(params=[param], weight_decay=training_args.weight_decay, **galore_kwargs)] optimizer_dict[param] = optim_class(param_groups, **optim_kwargs) def optimizer_hook(param: "torch.nn.Parameter"): if param.grad is not None: optimizer_dict[param].step() optimizer_dict[param].zero_grad() for param in trainable_params: param.register_post_accumulate_grad_hook(optimizer_hook) optimizer = DummyOptimizer(lr=training_args.learning_rate, optimizer_dict=optimizer_dict) else: param_groups = [ dict(params=nodecay_params, weight_decay=0.0), dict(params=decay_params, weight_decay=training_args.weight_decay), dict(params=galore_params, weight_decay=training_args.weight_decay, **galore_kwargs), ] optimizer = optim_class(param_groups, **optim_kwargs) logger.info_rank0( f"Using GaLore optimizer with args: {galore_kwargs}. " "It may cause hanging at the start of training, wait patiently." ) return optimizer def _create_apollo_optimizer( model: "PreTrainedModel", training_args: "TrainingArguments", finetuning_args: "FinetuningArguments", ) -> "torch.optim.Optimizer": if len(finetuning_args.apollo_target) == 1 and finetuning_args.apollo_target[0] == "all": apollo_targets = find_all_linear_modules(model, finetuning_args.freeze_vision_tower) else: apollo_targets = finetuning_args.apollo_target apollo_params: list[torch.nn.Parameter] = [] for name, module in model.named_modules(): if isinstance(module, torch.nn.Linear) and any(target in name for target in apollo_targets): for param in module.parameters(): if param.requires_grad and len(param.shape) > 1: apollo_params.append(param) apollo_kwargs = { "rank": finetuning_args.apollo_rank, "proj": finetuning_args.apollo_proj, "proj_type": finetuning_args.apollo_proj_type, "update_proj_gap": finetuning_args.apollo_update_interval, "scale": finetuning_args.apollo_scale, "scale_type": finetuning_args.apollo_scale_type, "scale_front": finetuning_args.apollo_scale_front, } id_apollo_params = {id(param) for param in apollo_params} decay_params, nodecay_params = [], [] # they are non-apollo parameters trainable_params: list[torch.nn.Parameter] = [] # apollo_params + decay_params + nodecay_params decay_param_names = _get_decay_parameter_names(model) for name, param in model.named_parameters(): if param.requires_grad: trainable_params.append(param) if id(param) not in id_apollo_params: if name in decay_param_names: decay_params.append(param) else: nodecay_params.append(param) _, optim_kwargs = Trainer.get_optimizer_cls_and_kwargs(training_args) if training_args.optim == "adamw_torch": optim_class = APOLLOAdamW else: raise NotImplementedError(f"Unknown optim: {training_args.optim}.") if finetuning_args.apollo_layerwise: logger.warning_rank0("The displayed gradient norm will be all zeros in layerwise APOLLO.") if training_args.gradient_accumulation_steps != 1: raise ValueError("Per-layer APOLLO does not support gradient accumulation.") optimizer_dict: dict[torch.Tensor, torch.optim.Optimizer] = {} for param in nodecay_params: param_groups = [dict(params=[param], weight_decay=0.0)] optimizer_dict[param] = optim_class(param_groups, **optim_kwargs) for param in decay_params: param_groups = [dict(params=[param], weight_decay=training_args.weight_decay)] optimizer_dict[param] = optim_class(param_groups, **optim_kwargs) for param in apollo_params: # apollo params have weight decay param_groups = [dict(params=[param], weight_decay=training_args.weight_decay, **apollo_kwargs)] optimizer_dict[param] = optim_class(param_groups, **optim_kwargs) def optimizer_hook(param: "torch.nn.Parameter"): if param.grad is not None: optimizer_dict[param].step() optimizer_dict[param].zero_grad() for param in trainable_params: param.register_post_accumulate_grad_hook(optimizer_hook) optimizer = DummyOptimizer(lr=training_args.learning_rate, optimizer_dict=optimizer_dict) else: param_groups = [ dict(params=nodecay_params, weight_decay=0.0), dict(params=decay_params, weight_decay=training_args.weight_decay), dict(params=apollo_params, weight_decay=training_args.weight_decay, **apollo_kwargs), ] optimizer = optim_class(param_groups, **optim_kwargs) logger.info_rank0(f"Using APOLLO optimizer with args: {apollo_kwargs}.") return optimizer def _create_loraplus_optimizer( model: "PreTrainedModel", training_args: "TrainingArguments", finetuning_args: "FinetuningArguments", ) -> "torch.optim.Optimizer": default_lr = training_args.learning_rate loraplus_lr = training_args.learning_rate * finetuning_args.loraplus_lr_ratio embedding_lr = finetuning_args.loraplus_lr_embedding decay_param_names = _get_decay_parameter_names(model) param_dict: dict[str, list[torch.nn.Parameter]] = { "lora_a": [], "lora_b": [], "lora_b_nodecay": [], "embedding": [], } for name, param in model.named_parameters(): if param.requires_grad: if "lora_embedding_B" in name: param_dict["embedding"].append(param) elif "lora_B" in name or param.ndim == 1: if name in decay_param_names: param_dict["lora_b"].append(param) else: param_dict["lora_b_nodecay"].append(param) else: param_dict["lora_a"].append(param) optim_class, optim_kwargs = Trainer.get_optimizer_cls_and_kwargs(training_args) param_groups = [ dict(params=param_dict["lora_a"], lr=default_lr, weight_decay=training_args.weight_decay), dict(params=param_dict["lora_b"], lr=loraplus_lr, weight_decay=training_args.weight_decay), dict(params=param_dict["lora_b_nodecay"], lr=loraplus_lr, weight_decay=0.0), dict(params=param_dict["embedding"], lr=embedding_lr, weight_decay=training_args.weight_decay), ] optimizer = optim_class(param_groups, **optim_kwargs) logger.info_rank0(f"Using LoRA+ optimizer with loraplus lr ratio {finetuning_args.loraplus_lr_ratio:.2f}.") return optimizer def _create_badam_optimizer( model: "PreTrainedModel", training_args: "TrainingArguments", finetuning_args: "FinetuningArguments", ) -> "torch.optim.Optimizer": decay_params, nodecay_params = [], [] decay_param_names = _get_decay_parameter_names(model) for name, param in model.named_parameters(): if param.requires_grad: if name in decay_param_names: decay_params.append(param) else: nodecay_params.append(param) optim_class, optim_kwargs = Trainer.get_optimizer_cls_and_kwargs(training_args) param_groups = [ dict(params=nodecay_params, weight_decay=0.0), dict(params=decay_params, weight_decay=training_args.weight_decay), ] if finetuning_args.badam_mode == "layer": from badam import BlockOptimizer # type: ignore base_optimizer = optim_class(param_groups, **optim_kwargs) optimizer = BlockOptimizer( base_optimizer=base_optimizer, named_parameters_list=list(model.named_parameters()), block_prefix_list=None, switch_block_every=finetuning_args.badam_switch_interval, start_block=finetuning_args.badam_start_block, switch_mode=finetuning_args.badam_switch_mode, verbose=finetuning_args.badam_verbose, ds_zero3_enabled=is_deepspeed_zero3_enabled(), ) logger.info_rank0( f"Using BAdam optimizer with layer-wise update, switch mode is {finetuning_args.badam_switch_mode}, " f"switch block every {finetuning_args.badam_switch_interval} steps, " f"default start block is {finetuning_args.badam_start_block}" ) elif finetuning_args.badam_mode == "ratio": from badam import BlockOptimizerRatio # type: ignore assert finetuning_args.badam_update_ratio > 1e-6 optimizer = BlockOptimizerRatio( param_groups=param_groups, named_parameters_list=list(model.named_parameters()), update_ratio=finetuning_args.badam_update_ratio, mask_mode=finetuning_args.badam_mask_mode, verbose=finetuning_args.badam_verbose, include_embedding=False, **optim_kwargs, ) logger.info_rank0( f"Using BAdam optimizer with ratio-based update, update ratio is {finetuning_args.badam_update_ratio}, " f"mask mode is {finetuning_args.badam_mask_mode}" ) return optimizer def _create_adam_mini_optimizer( model: "PreTrainedModel", training_args: "TrainingArguments", ) -> "torch.optim.Optimizer": from adam_mini import Adam_mini # type: ignore hidden_size = getattr(model.config, "hidden_size", None) num_q_head = getattr(model.config, "num_attention_heads", None) num_kv_head = getattr(model.config, "num_key_value_heads", None) optimizer = Adam_mini( named_parameters=model.named_parameters(), lr=training_args.learning_rate, betas=(training_args.adam_beta1, training_args.adam_beta2), eps=training_args.adam_epsilon, weight_decay=training_args.weight_decay, model_sharding=is_fsdp_enabled() or is_deepspeed_zero3_enabled(), dim=hidden_size, n_heads=num_q_head, n_kv_heads=num_kv_head, ) logger.info_rank0("Using Adam-mini optimizer.") return optimizer def _create_muon_optimizer( model: "PreTrainedModel", training_args: "TrainingArguments", ) -> "torch.optim.Optimizer": from ..third_party.muon import Muon muon_params, adamw_params = [], [] for name, param in model.named_parameters(): if param.requires_grad: # Use Muon for 2D parameters that aren't embeddings or heads if param.ndim == 2 and "embed" not in name and "lm_head" not in name: muon_params.append(param) else: adamw_params.append(param) optimizer = Muon( lr=training_args.learning_rate, wd=training_args.weight_decay, muon_params=muon_params, adamw_params=adamw_params, adamw_betas=(training_args.adam_beta1, training_args.adam_beta2), adamw_eps=training_args.adam_epsilon, ) logger.info_rank0( f"Using Muon optimizer with {len(muon_params)} Muon params and {len(adamw_params)} AdamW params." ) return optimizer def create_custom_optimizer( model: "PreTrainedModel", training_args: "TrainingArguments", finetuning_args: "FinetuningArguments", ) -> Optional["torch.optim.Optimizer"]: if finetuning_args.use_galore: return _create_galore_optimizer(model, training_args, finetuning_args) if finetuning_args.use_apollo: return _create_apollo_optimizer(model, training_args, finetuning_args) if finetuning_args.loraplus_lr_ratio is not None: return _create_loraplus_optimizer(model, training_args, finetuning_args) if finetuning_args.use_badam: return _create_badam_optimizer(model, training_args, finetuning_args) if finetuning_args.use_adam_mini: return _create_adam_mini_optimizer(model, training_args) if finetuning_args.use_muon: return _create_muon_optimizer(model, training_args) def create_custom_scheduler( training_args: "TrainingArguments", num_training_steps: int, optimizer: Optional["torch.optim.Optimizer"] = None, ) -> None: if training_args.lr_scheduler_type == "warmup_stable_decay": num_warmup_steps = training_args.get_warmup_steps(num_training_steps) remaining_steps = num_training_steps - num_warmup_steps num_stable_steps = remaining_steps // 3 # use 1/3 for stable by default num_decay_steps = remaining_steps - num_stable_steps scheduler_kwargs = training_args.lr_scheduler_kwargs or {} default_kwargs = { "num_stable_steps": num_stable_steps, "num_decay_steps": num_decay_steps, } for key, value in default_kwargs.items(): if key not in scheduler_kwargs: scheduler_kwargs[key] = value training_args.lr_scheduler_kwargs = scheduler_kwargs if optimizer is not None and isinstance(optimizer, DummyOptimizer): optimizer_dict = optimizer.optimizer_dict scheduler_dict: dict[torch.nn.Parameter, torch.optim.lr_scheduler.LRScheduler] = {} for param in optimizer_dict.keys(): scheduler_dict[param] = get_scheduler( training_args.lr_scheduler_type, optimizer=optimizer_dict[param], num_warmup_steps=training_args.get_warmup_steps(num_training_steps), num_training_steps=num_training_steps, scheduler_specific_kwargs=training_args.lr_scheduler_kwargs, ) def scheduler_hook(param: "torch.nn.Parameter"): scheduler_dict[param].step() for param in optimizer_dict.keys(): param.register_post_accumulate_grad_hook(scheduler_hook) def get_batch_logps( logits: "torch.Tensor", labels: "torch.Tensor", label_pad_token_id: int = IGNORE_INDEX, ld_alpha: Optional[float] = None, ) -> tuple["torch.Tensor", "torch.Tensor"]: r"""Compute the log probabilities of the given labels under the given logits. Returns: logps: A tensor of shape (batch_size,) containing the sum of log probabilities. valid_length: A tensor of shape (batch_size,) containing the number of non-masked tokens. """ if logits.shape[:-1] != labels.shape: raise ValueError("Logits (batchsize x seqlen) and labels must have the same shape.") labels = labels[:, 1:].clone() logits = logits[:, :-1, :] loss_mask = labels != label_pad_token_id labels[labels == label_pad_token_id] = 0 # dummy token per_token_logps = torch.gather(logits.log_softmax(-1), dim=2, index=labels.unsqueeze(2)).squeeze(2) valid_length = loss_mask.sum(-1) if ld_alpha is not None: num_examples = labels.shape[0] // 2 chosen_lengths = valid_length[:num_examples] rejected_lengths = valid_length[num_examples:] min_lengths = torch.min(chosen_lengths, rejected_lengths) start_positions = torch.argmax(loss_mask.int(), dim=1) public_lengths = start_positions + torch.cat([min_lengths, min_lengths], dim=0) seq_len = labels.shape[-1] position_ids = torch.arange(seq_len, device=per_token_logps.device).expand_as(per_token_logps) ld_mask = position_ids < public_lengths.unsqueeze(1) front_mask = (ld_mask * loss_mask).float() rear_mask = (~ld_mask * loss_mask).float() front_logps = (per_token_logps * front_mask).sum(-1) rear_logps = (per_token_logps * rear_mask).sum(-1) logps = front_logps + ld_alpha * rear_logps else: logps = (per_token_logps * loss_mask).sum(-1) return logps, valid_length def dft_loss_func(outputs, labels, num_items_in_batch=None): logits = outputs.get("logits") if logits is None: return outputs.get("loss", torch.tensor(0.0)) logits = logits.float() vocab_size = logits.size(-1) labels = torch.nn.functional.pad(labels, (0, 1), value=-100) shift_labels = labels[..., 1:].contiguous() logits = logits.view(-1, vocab_size) shift_labels = shift_labels.view(-1) shift_labels = shift_labels.to(logits.device) loss = _dft_cross_entropy(logits, shift_labels, num_items_in_batch) return loss def _dft_cross_entropy( source: torch.Tensor, target: torch.Tensor, num_items_in_batch: Optional[torch.Tensor] = None, ignore_index: int = -100, ) -> torch.Tensor: per_token_loss = torch.nn.functional.cross_entropy(source, target, ignore_index=ignore_index, reduction="none") valid_mask = target != ignore_index if not valid_mask.any(): return torch.tensor(0.0, device=source.device, dtype=source.dtype) valid_losses = per_token_loss[valid_mask] with torch.no_grad(): target_probs = torch.exp(-valid_losses) weighted_losses = valid_losses * target_probs if num_items_in_batch is not None: total_loss = weighted_losses.sum() if torch.is_tensor(num_items_in_batch): num_items_in_batch = num_items_in_batch.to(total_loss.device) loss = total_loss / num_items_in_batch else: loss = weighted_losses.mean() return loss def nested_detach( tensors: Union["torch.Tensor", list["torch.Tensor"], tuple["torch.Tensor"], dict[str, "torch.Tensor"]], clone: bool = False, ): r"""Detach `tensors` (even if it's a nested list/tuple/dict of tensors).""" if isinstance(tensors, (list, tuple)): return type(tensors)(nested_detach(t, clone=clone) for t in tensors) elif isinstance(tensors, Mapping): return type(tensors)({k: nested_detach(t, clone=clone) for k, t in tensors.items()}) if isinstance(tensors, torch.Tensor): if clone: return tensors.detach().clone() else: return tensors.detach() else: return tensors def get_swanlab_callback(finetuning_args: "FinetuningArguments") -> "TrainerCallback": r"""Get the callback for logging to SwanLab.""" import swanlab # type: ignore from swanlab.integration.transformers import SwanLabCallback # type: ignore if finetuning_args.swanlab_api_key is not None: swanlab.login(api_key=finetuning_args.swanlab_api_key) if finetuning_args.swanlab_lark_webhook_url is not None: from swanlab.plugin.notification import LarkCallback # type: ignore lark_callback = LarkCallback( webhook_url=finetuning_args.swanlab_lark_webhook_url, secret=finetuning_args.swanlab_lark_secret, ) swanlab.register_callbacks([lark_callback]) class SwanLabCallbackExtension(SwanLabCallback): def setup(self, args: "TrainingArguments", state: "TrainerState", model: "PreTrainedModel", **kwargs): if not state.is_world_process_zero: return super().setup(args, state, model, **kwargs) try: if hasattr(self, "_swanlab"): swanlab_public_config = self._swanlab.get_run().public.json() else: # swanlab <= 0.4.9 swanlab_public_config = self._experiment.get_run().public.json() except Exception: swanlab_public_config = {} with open(os.path.join(args.output_dir, SWANLAB_CONFIG), "w") as f: f.write(json.dumps(swanlab_public_config, indent=2)) swanlab_callback = SwanLabCallbackExtension( project=finetuning_args.swanlab_project, workspace=finetuning_args.swanlab_workspace, experiment_name=finetuning_args.swanlab_run_name, mode=finetuning_args.swanlab_mode, config={"Framework": "πŸ¦™LlamaFactory"}, logdir=finetuning_args.swanlab_logdir, tags=["πŸ¦™LlamaFactory"], ) return swanlab_callback def get_ray_trainer( training_function: Callable, train_loop_config: dict[str, Any], ray_args: "RayArguments", ) -> "TorchTrainer": if not ray_args.use_ray: raise ValueError("Ray was not enabled. Please set `USE_RAY=1` to enable ray.") if ray_args.ray_init_kwargs is not None: ray.init(**ray_args.ray_init_kwargs) if ray_args.ray_storage_filesystem is not None: # this means we are using s3/gcs storage_path = ray_args.ray_storage_path else: storage_path = Path(ray_args.ray_storage_path).absolute().as_posix() trainer = TorchTrainer( training_function, train_loop_config=train_loop_config, scaling_config=ScalingConfig( num_workers=ray_args.ray_num_workers, resources_per_worker=ray_args.resources_per_worker, placement_strategy=ray_args.placement_strategy, use_gpu=True, ), run_config=RunConfig( name=ray_args.ray_run_name, storage_filesystem=ray_args.ray_storage_filesystem, storage_path=storage_path, ), ) return trainer
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/train/test_utils.py
src/llamafactory/train/test_utils.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING, Optional, Union import torch from peft import PeftModel from transformers import AutoModelForCausalLM from trl import AutoModelForCausalLMWithValueHead from ..data import get_dataset, get_template_and_fix_tokenizer from ..hparams import get_infer_args, get_train_args from ..model import load_model, load_tokenizer if TYPE_CHECKING: from peft import LoraModel from transformers import PreTrainedModel from ..data.data_utils import DatasetModule def compare_model(model_a: "torch.nn.Module", model_b: "torch.nn.Module", diff_keys: list[str] = []) -> None: state_dict_a = model_a.state_dict() state_dict_b = model_b.state_dict() assert set(state_dict_a.keys()) == set(state_dict_b.keys()) for name in state_dict_a.keys(): if any(key in name for key in diff_keys): assert torch.allclose(state_dict_a[name], state_dict_b[name], rtol=1e-4, atol=1e-5) is False else: assert torch.allclose(state_dict_a[name], state_dict_b[name], rtol=1e-4, atol=1e-5) is True def check_lora_model(model: "LoraModel") -> tuple[set[str], set[str]]: linear_modules, extra_modules = set(), set() for name, param in model.named_parameters(): if any(module in name for module in ["lora_A", "lora_B"]): linear_modules.add(name.split(".lora_", maxsplit=1)[0].split(".")[-1]) assert param.requires_grad is True assert param.dtype == torch.float32 elif "modules_to_save" in name: extra_modules.add(name.split(".modules_to_save", maxsplit=1)[0].split(".")[-1]) assert param.requires_grad is True assert param.dtype == torch.float32 else: assert param.requires_grad is False assert param.dtype == torch.float16 return linear_modules, extra_modules def load_train_model(add_valuehead: bool = False, **kwargs) -> "PreTrainedModel": model_args, _, _, finetuning_args, _ = get_train_args(kwargs) tokenizer = load_tokenizer(model_args)["tokenizer"] return load_model(tokenizer, model_args, finetuning_args, is_trainable=True, add_valuehead=add_valuehead) def load_infer_model(add_valuehead: bool = False, **kwargs) -> "PreTrainedModel": model_args, _, finetuning_args, _ = get_infer_args(kwargs) tokenizer = load_tokenizer(model_args)["tokenizer"] return load_model(tokenizer, model_args, finetuning_args, is_trainable=False, add_valuehead=add_valuehead) def load_reference_model( model_path: str, lora_path: Optional[str] = None, use_lora: bool = False, use_pissa: bool = False, is_trainable: bool = False, add_valuehead: bool = False, ) -> Union["PreTrainedModel", "LoraModel"]: if add_valuehead: model: AutoModelForCausalLMWithValueHead = AutoModelForCausalLMWithValueHead.from_pretrained( model_path, torch_dtype=torch.float16, device_map="auto" ) return model model = AutoModelForCausalLM.from_pretrained(model_path, torch_dtype=torch.float16, device_map="auto") if use_lora or use_pissa: model = PeftModel.from_pretrained( model, lora_path, subfolder="pissa_init" if use_pissa else None, is_trainable=is_trainable ) for param in filter(lambda p: p.requires_grad, model.parameters()): param.data = param.data.to(torch.float32) return model def load_dataset_module(**kwargs) -> "DatasetModule": model_args, data_args, training_args, _, _ = get_train_args(kwargs) tokenizer_module = load_tokenizer(model_args) template = get_template_and_fix_tokenizer(tokenizer_module["tokenizer"], data_args) dataset_module = get_dataset(template, model_args, data_args, training_args, kwargs["stage"], **tokenizer_module) return dataset_module def patch_valuehead_model() -> None: def post_init(self: "AutoModelForCausalLMWithValueHead", state_dict: dict[str, "torch.Tensor"]) -> None: state_dict = {k[7:]: state_dict[k] for k in state_dict.keys() if k.startswith("v_head.")} self.v_head.load_state_dict(state_dict, strict=False) del state_dict AutoModelForCausalLMWithValueHead.post_init = post_init
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/train/tuner.py
src/llamafactory/train/tuner.py
# Copyright 2025 the KVCache.AI team, Approaching AI, and the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import os import shutil from typing import TYPE_CHECKING, Any, Optional import torch import torch.distributed as dist from transformers import EarlyStoppingCallback, PreTrainedModel from ..data import get_template_and_fix_tokenizer from ..extras import logging from ..extras.constants import V_HEAD_SAFE_WEIGHTS_NAME, V_HEAD_WEIGHTS_NAME from ..extras.misc import infer_optim_dtype from ..extras.packages import is_mcore_adapter_available, is_ray_available from ..hparams import get_infer_args, get_ray_args, get_train_args, read_args from ..model import load_model, load_tokenizer from .callbacks import LogCallback, PissaConvertCallback, ReporterCallback from .dpo import run_dpo from .kto import run_kto from .ppo import run_ppo from .pt import run_pt from .rm import run_rm from .sft import run_sft from .trainer_utils import get_ray_trainer, get_swanlab_callback if is_ray_available(): import ray from ray.train.huggingface.transformers import RayTrainReportCallback if TYPE_CHECKING: from transformers import TrainerCallback logger = logging.get_logger(__name__) def _training_function(config: dict[str, Any]) -> None: args = config.get("args") callbacks: list[Any] = config.get("callbacks") model_args, data_args, training_args, finetuning_args, generating_args = get_train_args(args) callbacks.append(LogCallback()) if finetuning_args.pissa_convert: callbacks.append(PissaConvertCallback()) if finetuning_args.use_swanlab: callbacks.append(get_swanlab_callback(finetuning_args)) if finetuning_args.early_stopping_steps is not None: callbacks.append(EarlyStoppingCallback(early_stopping_patience=finetuning_args.early_stopping_steps)) callbacks.append(ReporterCallback(model_args, data_args, finetuning_args, generating_args)) # add to last if finetuning_args.stage in ["pt", "sft", "dpo"] and finetuning_args.use_mca: if not is_mcore_adapter_available(): raise ImportError("mcore_adapter is not installed. Please install it with `pip install mcore-adapter`.") if finetuning_args.stage == "pt": from .mca import run_pt as run_pt_mca run_pt_mca(model_args, data_args, training_args, finetuning_args, callbacks) elif finetuning_args.stage == "sft": from .mca import run_sft as run_sft_mca run_sft_mca(model_args, data_args, training_args, finetuning_args, callbacks) elif finetuning_args.stage == "dpo": from .mca import run_dpo as run_dpo_mca run_dpo_mca(model_args, data_args, training_args, finetuning_args, callbacks) elif finetuning_args.stage == "pt": run_pt(model_args, data_args, training_args, finetuning_args, callbacks) elif finetuning_args.stage == "sft": run_sft(model_args, data_args, training_args, finetuning_args, generating_args, callbacks) elif finetuning_args.stage == "rm": run_rm(model_args, data_args, training_args, finetuning_args, callbacks) elif finetuning_args.stage == "ppo": run_ppo(model_args, data_args, training_args, finetuning_args, generating_args, callbacks) elif finetuning_args.stage == "dpo": run_dpo(model_args, data_args, training_args, finetuning_args, callbacks) elif finetuning_args.stage == "kto": run_kto(model_args, data_args, training_args, finetuning_args, callbacks) else: raise ValueError(f"Unknown task: {finetuning_args.stage}.") if is_ray_available() and ray.is_initialized(): return # if ray is intialized it will destroy the process group on return try: if dist.is_initialized(): dist.destroy_process_group() except Exception as e: logger.warning(f"Failed to destroy process group: {e}.") def run_exp(args: Optional[dict[str, Any]] = None, callbacks: Optional[list["TrainerCallback"]] = None) -> None: args = read_args(args) if "-h" in args or "--help" in args: get_train_args(args) ray_args = get_ray_args(args) callbacks = callbacks or [] if ray_args.use_ray: callbacks.append(RayTrainReportCallback()) trainer = get_ray_trainer( training_function=_training_function, train_loop_config={"args": args, "callbacks": callbacks}, ray_args=ray_args, ) trainer.fit() else: _training_function(config={"args": args, "callbacks": callbacks}) def export_model(args: Optional[dict[str, Any]] = None) -> None: model_args, data_args, finetuning_args, _ = get_infer_args(args) if model_args.export_dir is None: raise ValueError("Please specify `export_dir` to save model.") if model_args.adapter_name_or_path is not None and model_args.export_quantization_bit is not None: raise ValueError("Please merge adapters before quantizing the model.") tokenizer_module = load_tokenizer(model_args) tokenizer = tokenizer_module["tokenizer"] processor = tokenizer_module["processor"] template = get_template_and_fix_tokenizer(tokenizer, data_args) model = load_model(tokenizer, model_args, finetuning_args) # must after fixing tokenizer to resize vocab if getattr(model, "quantization_method", None) is not None and model_args.adapter_name_or_path is not None: raise ValueError("Cannot merge adapters to a quantized model.") if not isinstance(model, PreTrainedModel): raise ValueError("The model is not a `PreTrainedModel`, export aborted.") if getattr(model, "quantization_method", None) is not None: # quantized model adopts float16 type setattr(model.config, "torch_dtype", torch.float16) else: if model_args.infer_dtype == "auto": output_dtype = getattr(model.config, "torch_dtype", torch.float32) if output_dtype == torch.float32: # if infer_dtype is auto, try using half precision first output_dtype = infer_optim_dtype(torch.bfloat16) else: output_dtype = getattr(torch, model_args.infer_dtype) setattr(model.config, "torch_dtype", output_dtype) model = model.to(output_dtype) logger.info_rank0(f"Convert model dtype to: {output_dtype}.") model.save_pretrained( save_directory=model_args.export_dir, max_shard_size=f"{model_args.export_size}GB", safe_serialization=(not model_args.export_legacy_format), ) if model_args.export_hub_model_id is not None: model.push_to_hub( model_args.export_hub_model_id, token=model_args.hf_hub_token, max_shard_size=f"{model_args.export_size}GB", safe_serialization=(not model_args.export_legacy_format), ) if finetuning_args.stage == "rm": if model_args.adapter_name_or_path is not None: vhead_path = model_args.adapter_name_or_path[-1] else: vhead_path = model_args.model_name_or_path if os.path.exists(os.path.join(vhead_path, V_HEAD_SAFE_WEIGHTS_NAME)): shutil.copy( os.path.join(vhead_path, V_HEAD_SAFE_WEIGHTS_NAME), os.path.join(model_args.export_dir, V_HEAD_SAFE_WEIGHTS_NAME), ) logger.info_rank0(f"Copied valuehead to {model_args.export_dir}.") elif os.path.exists(os.path.join(vhead_path, V_HEAD_WEIGHTS_NAME)): shutil.copy( os.path.join(vhead_path, V_HEAD_WEIGHTS_NAME), os.path.join(model_args.export_dir, V_HEAD_WEIGHTS_NAME), ) logger.info_rank0(f"Copied valuehead to {model_args.export_dir}.") try: tokenizer.padding_side = "left" # restore padding side tokenizer.init_kwargs["padding_side"] = "left" tokenizer.save_pretrained(model_args.export_dir) if model_args.export_hub_model_id is not None: tokenizer.push_to_hub(model_args.export_hub_model_id, token=model_args.hf_hub_token) if processor is not None: processor.save_pretrained(model_args.export_dir) if model_args.export_hub_model_id is not None: processor.push_to_hub(model_args.export_hub_model_id, token=model_args.hf_hub_token) except Exception as e: logger.warning_rank0(f"Cannot save tokenizer, please copy the files manually: {e}.") ollama_modelfile = os.path.join(model_args.export_dir, "Modelfile") with open(ollama_modelfile, "w", encoding="utf-8") as f: f.write(template.get_ollama_modelfile(tokenizer)) logger.info_rank0(f"Ollama modelfile saved in {ollama_modelfile}")
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/train/__init__.py
src/llamafactory/train/__init__.py
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/train/fp8_utils.py
src/llamafactory/train/fp8_utils.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import os import types from typing import TYPE_CHECKING, Any, Optional from ..extras import logging if TYPE_CHECKING: from ..hparams import TrainingArguments logger = logging.get_logger(__name__) def create_fp8_kwargs(training_args: "TrainingArguments") -> list[Any]: """Create AORecipeKwargs for FP8 training with HuggingFace Accelerate. Args: training_args: Training arguments containing FP8 configuration Returns: List containing AORecipeKwargs if FP8 is enabled and supported, empty list otherwise """ if not training_args.fp8: return [] backend = getattr(training_args, "fp8_backend", "auto") logger.info_rank0(f"Creating FP8 configuration with backend: {backend}") try: # Use Transformer Engine backend (optimal for Hopper GPUs) if backend == "te": from accelerate.utils import FP8RecipeKwargs logger.info_rank0("Using Transformer Engine FP8 backend") return [FP8RecipeKwargs(backend="TE", fp8_format="HYBRID", amax_history_len=16, amax_compute_algo="max")] # Use TorchAO backend (default) from accelerate.utils import AORecipeKwargs # Create Float8LinearConfig if torchao backend is used config = None if backend == "torchao" or backend == "auto": from torchao.float8 import Float8LinearConfig # Use rowwise scaling for better performance (as recommended by torchao) # Configure alignment requirements for FP8 kernels config = Float8LinearConfig.from_recipe_name("rowwise") # Enable alignment for better kernel performance if hasattr(config, "enable_amax_init"): config.enable_amax_init = True if hasattr(config, "enable_pre_and_post_forward"): config.enable_pre_and_post_forward = True # Create module filter function to skip problematic layers # TorchAO FP8 requires dimensions divisible by 16 for optimal kernels def module_filter_func(module, layer_name): # Skip embedding and output layers for numerical stability skip_layers = ["embed", "lm_head", "output", "classifier"] if any(skip_name in layer_name.lower() for skip_name in skip_layers): return False # Only convert Linear layers if not (hasattr(module, "weight") and len(module.weight.shape) == 2): return False # Check dimension alignment for FP8 kernels weight = module.weight in_features, out_features = weight.shape[1], weight.shape[0] # Skip layers with dimensions not divisible by 16 to avoid kernel errors if in_features % 16 != 0 or out_features % 16 != 0: logger.debug( f"Skipping layer {layer_name} with dimensions {out_features}x{in_features} (not divisible by 16)" ) return False return True # Map FSDP all-gather setting if available (this affects the underlying implementation) if ( hasattr(training_args, "fp8_enable_fsdp_float8_all_gather") and training_args.fp8_enable_fsdp_float8_all_gather ): logger.info_rank0("FSDP float8 all-gather optimization requested") return [AORecipeKwargs(config=config, module_filter_func=module_filter_func)] except Exception as e: logger.info_rank0(f"Failed to create FP8 configuration: {e}") return [] def get_fp8_mixed_precision(training_args: "TrainingArguments") -> Optional[str]: """Get the mixed precision setting for Accelerate when using FP8. Args: training_args: Training arguments containing FP8 configuration Returns: "fp8" if FP8 is enabled, None otherwise """ return "fp8" if training_args.fp8 else None def configure_fp8_environment(training_args: "TrainingArguments") -> None: """Configure FP8 environment for HuggingFace Accelerate. FP8 training is handled entirely through HuggingFace Accelerate, regardless of whether DeepSpeed or FSDP is used for distributed training. This function sets up the environment variables and validates the FP8 configuration. Args: training_args: Training arguments containing FP8 configuration """ if not training_args.fp8: return # Set mixed precision to fp8 for HuggingFace Accelerate os.environ["ACCELERATE_MIXED_PRECISION"] = "fp8" logger.info_rank0("Set ACCELERATE_MIXED_PRECISION=fp8") # Configure FP8 backend and options backend = getattr(training_args, "fp8_backend", "auto") if backend != "auto": os.environ["FP8_BACKEND"] = backend logger.info_rank0(f"Set FP8_BACKEND={backend}") # Create and validate FP8 recipe kwargs (for logging/debugging) fp8_kwargs = create_fp8_kwargs(training_args) logger.info_rank0(f"FP8 AORecipeKwargs created: {len(fp8_kwargs)} items") # Enable FSDP float8 all-gather optimization if requested if hasattr(training_args, "fp8_enable_fsdp_float8_all_gather") and training_args.fp8_enable_fsdp_float8_all_gather: os.environ["FP8_ENABLE_FSDP_FLOAT8_ALL_GATHER"] = "true" logger.info_rank0("Set FP8_ENABLE_FSDP_FLOAT8_ALL_GATHER=true") logger.info_rank0("FP8 environment configured - all FP8 training handled by HuggingFace Accelerate") def verify_fp8_status(accelerator, training_args: "TrainingArguments") -> None: """Verify that FP8 training is actually working after model preparation. Args: accelerator: The HuggingFace Accelerator instance training_args: Training arguments containing FP8 configuration """ if not training_args.fp8: return # Check Accelerate's FP8 status fp8_enabled = getattr(accelerator, "fp8_enabled", False) fp8_backend_type = getattr(accelerator, "fp8_backend", "UNKNOWN") backend = getattr(training_args, "fp8_backend", "auto") if backend == "torchao" or backend == "auto": logger.info_rank0( "FP8 training enabled with TorchAO backend. For optimal performance, " "ensure model layer dimensions are mostly divisible by 16. " "If you encounter issues, try fp8_backend='te' with Transformer Engine." ) else: logger.info_rank0(f"FP8 training enabled with {backend} backend.") logger.info_rank0(f"Accelerate FP8 status - enabled: {fp8_enabled}, backend: {fp8_backend_type}") if not fp8_enabled: logger.info_rank0("WARNING: FP8 was requested but Accelerate shows fp8_enabled=False. FP8 may not be working.") def patch_accelerator_for_fp8() -> None: """Patch Accelerator to inject FP8 recipe kwargs. This is needed because HuggingFace Trainer doesn't pass kwargs_handlers to Accelerator. We monkey-patch Accelerator.__init__ to inject the FP8 recipe and force mixed_precision='fp8'. """ import transformer_engine.pytorch as te from accelerate import Accelerator # Guard against multiple patches if getattr(Accelerator, "_te_fp8_patched", False): return # Stub for Accelerate 1.12+ compatibility (te.fp8.check_mxfp8_support doesn't exist yet) if not hasattr(te, "fp8"): te.fp8 = types.ModuleType("fp8") te.fp8.check_mxfp8_support = lambda: (False, "MXFP8 not supported") try: from accelerate.utils import TERecipeKwargs as FP8Recipe use_te_recipe = True except ImportError: from accelerate.utils import FP8RecipeKwargs as FP8Recipe use_te_recipe = False original_init = Accelerator.__init__ def patched_init(self, *args, **kwargs): if "kwargs_handlers" not in kwargs or not kwargs["kwargs_handlers"]: if use_te_recipe: kwargs["kwargs_handlers"] = [ FP8Recipe(fp8_format="HYBRID", amax_history_len=16, amax_compute_algo="max") ] else: kwargs["kwargs_handlers"] = [ FP8Recipe(backend="TE", fp8_format="HYBRID", amax_history_len=16, amax_compute_algo="max") ] # Only force mixed_precision when we inject handlers kwargs["mixed_precision"] = "fp8" return original_init(self, *args, **kwargs) Accelerator.__init__ = patched_init Accelerator._te_fp8_patched = True
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/train/ppo/trainer.py
src/llamafactory/train/ppo/trainer.py
# Copyright 2025 HuggingFace Inc. and the LlamaFactory team. # # This code is inspired by the HuggingFace's TRL library. # https://github.com/huggingface/trl/blob/v0.8.0/trl/trainer/ppo_trainer.py # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import math import os import sys import warnings from types import MethodType from typing import TYPE_CHECKING, Any, Optional import torch from accelerate.utils import DistributedDataParallelKwargs from tqdm import tqdm from transformers import GenerationConfig, Trainer, TrainerControl, TrainerState from transformers.optimization import get_scheduler from transformers.trainer import DEFAULT_CALLBACKS from transformers.trainer_callback import CallbackHandler from transformers.trainer_pt_utils import remove_dummy_checkpoint from transformers.trainer_utils import PREFIX_CHECKPOINT_DIR from transformers.utils import SAFE_WEIGHTS_NAME, WEIGHTS_NAME from trl import PPOConfig, PPOTrainer from trl import __version__ as trl_version from trl.models.utils import unwrap_model_for_generation from typing_extensions import override from ...extras import logging from ...extras.misc import AverageMeter, count_parameters, get_current_device, get_logits_processor, torch_gc from ..callbacks import FixValueHeadModelCallback, SaveProcessorCallback from ..trainer_utils import create_custom_optimizer, create_custom_scheduler from .ppo_utils import dump_layernorm, get_rewards_from_server, replace_model, restore_layernorm if TYPE_CHECKING: from datasets import Dataset from transformers import ( DataCollatorWithPadding, PreTrainedTokenizer, ProcessorMixin, Seq2SeqTrainingArguments, TrainerCallback, ) from trl import AutoModelForCausalLMWithValueHead from ...hparams import FinetuningArguments, GeneratingArguments, ModelArguments logger = logging.get_logger(__name__) class CustomPPOTrainer(PPOTrainer, Trainer): r"""Inherit PPOTrainer.""" def __init__( self, model_args: "ModelArguments", training_args: "Seq2SeqTrainingArguments", finetuning_args: "FinetuningArguments", generating_args: "GeneratingArguments", callbacks: Optional[list["TrainerCallback"]], model: "AutoModelForCausalLMWithValueHead", reward_model: Optional["AutoModelForCausalLMWithValueHead"], ref_model: Optional["AutoModelForCausalLMWithValueHead"], tokenizer: "PreTrainedTokenizer", processor: Optional["ProcessorMixin"], data_collator: "DataCollatorWithPadding", train_dataset: Optional["Dataset"] = None, eval_dataset: Optional["Dataset"] = None, ) -> None: if eval_dataset is not None: raise NotImplementedError("PPOTrainer does not support eval dataset yet.") # Check if TRL version is compatible (0.8.6 <= version <= 0.9.6) try: from transformers.utils.versions import require_version require_version( "trl>=0.8.6,<=0.9.6", "Incompatible TRL version detected. LLaMA-Factory ppo requires TRL version >=0.8.6,<=0.9.6. " f"Found version {trl_version}. Please install the correct version with: `pip install trl>=0.8.6,<=0.9.6`\n" "To fix: run `DISABLE_VERSION_CHECK=1 llamafactory-cli train example_ppo.yaml`\n", ) except ImportError as e: raise e backward_batch_size = training_args.per_device_train_batch_size * training_args.gradient_accumulation_steps ppo_config = PPOConfig( model_name=model_args.model_name_or_path, learning_rate=training_args.learning_rate, mini_batch_size=training_args.per_device_train_batch_size, batch_size=backward_batch_size * finetuning_args.ppo_buffer_size, gradient_accumulation_steps=training_args.gradient_accumulation_steps, ppo_epochs=finetuning_args.ppo_epochs, max_grad_norm=training_args.max_grad_norm, seed=training_args.seed, optimize_device_cache=True, target=finetuning_args.ppo_target, use_score_scaling=finetuning_args.ppo_score_norm, use_score_norm=finetuning_args.ppo_score_norm, whiten_rewards=finetuning_args.ppo_whiten_rewards, accelerator_kwargs={"step_scheduler_with_optimizer": False}, log_with=training_args.report_to[0] if training_args.report_to else None, project_kwargs={"logging_dir": training_args.logging_dir}, ) # Add deepspeed config if training_args.deepspeed_plugin is not None: ppo_config.accelerator_kwargs["kwargs_handlers"] = [ DistributedDataParallelKwargs(find_unused_parameters=training_args.ddp_find_unused_parameters) ] ppo_config.accelerator_kwargs["deepspeed_plugin"] = training_args.deepspeed_plugin if ppo_config.log_with is not None: logger.warning_rank0("PPOTrainer cannot use external logger when DeepSpeed is enabled.") ppo_config.log_with = None # Create optimizer and scheduler if training_args.max_steps > 0: num_training_steps = training_args.max_steps else: total_train_batch_size = backward_batch_size * finetuning_args.ppo_buffer_size * training_args.world_size num_training_steps = training_args.num_train_epochs * math.ceil( len(train_dataset) / total_train_batch_size ) optimizer = self.create_optimizer(model, training_args, finetuning_args) scheduler = self.create_scheduler(training_args, num_training_steps, optimizer) PPOTrainer.__init__( self, config=ppo_config, model=model, ref_model=ref_model, tokenizer=tokenizer, dataset=train_dataset, optimizer=optimizer, data_collator=data_collator, lr_scheduler=scheduler, ) self.args = training_args self.model_args = model_args self.finetuning_args = finetuning_args self.reward_model = reward_model self.current_device = get_current_device() # patch for deepspeed training self.generation_config = GenerationConfig( pad_token_id=self.tokenizer.pad_token_id, eos_token_id=[self.tokenizer.eos_token_id] + self.tokenizer.additional_special_tokens_ids, **generating_args.to_dict(), ) self.state = TrainerState() self.control = TrainerControl() self.is_deepspeed_enabled = getattr(self.accelerator.state, "deepspeed_plugin", None) is not None self.is_fsdp_enabled = getattr(self.accelerator.state, "fsdp_plugin", None) is not None callbacks = DEFAULT_CALLBACKS if callbacks is None else DEFAULT_CALLBACKS + callbacks self.callback_handler = CallbackHandler( callbacks, self.accelerator.unwrap_model(self.model), self.tokenizer, self.optimizer, self.lr_scheduler ) if self.args.max_steps > 0: logger.info_rank0("max_steps is given, it will override any value given in num_train_epochs") self.amp_context = torch.autocast(self.current_device.type) warnings.simplefilter("ignore") # remove gc warnings on ref model if finetuning_args.reward_model_type == "full": if self.is_deepspeed_enabled: if not ( getattr(reward_model.pretrained_model, "is_loaded_in_8bit", False) or getattr(reward_model.pretrained_model, "is_loaded_in_4bit", False) ): # quantized models are already set on the correct device self.reward_model = self._prepare_deepspeed(self.reward_model) else: self.reward_model = self.accelerator.prepare_model(self.reward_model, evaluation_mode=True) self.add_callback(FixValueHeadModelCallback) if processor is not None: self.add_callback(SaveProcessorCallback(processor)) if finetuning_args.use_badam: from badam import BAdamCallback, clip_grad_norm_old_version # type: ignore self.accelerator.clip_grad_norm_ = MethodType(clip_grad_norm_old_version, self.accelerator) self.add_callback(BAdamCallback) def ppo_train(self, resume_from_checkpoint: Optional[str] = None) -> None: r"""Implement training loop for the PPO stage, like _inner_training_loop() in Huggingface's Trainer.""" if resume_from_checkpoint is not None: raise ValueError("`resume_from_checkpoint` will be supported in the future version.") total_train_batch_size = ( self.args.per_device_train_batch_size * self.args.gradient_accumulation_steps * self.finetuning_args.ppo_buffer_size * self.args.world_size ) if self.args.max_steps > 0: num_examples = total_train_batch_size * self.args.max_steps num_train_epochs = sys.maxsize max_steps = self.args.max_steps steps_in_epoch = self.args.max_steps else: len_dataloader = len(self.dataloader) num_examples = len(self.dataset) num_train_epochs = self.args.num_train_epochs max_steps = math.ceil(num_train_epochs * len_dataloader) steps_in_epoch = len_dataloader self.state.max_steps = max_steps self.state.num_train_epochs = num_train_epochs self.state.is_local_process_zero = self.is_local_process_zero() self.state.is_world_process_zero = self.is_world_process_zero() logger.info_rank0("***** Running training *****") logger.info_rank0(f" Num examples = {num_examples:,}") logger.info_rank0(f" Num Epochs = {num_train_epochs:,}") logger.info_rank0(f" Instantaneous batch size per device = {self.args.per_device_train_batch_size:,}") logger.info_rank0( f" Total train batch size (w. parallel, buffer, distributed & accumulation) = {total_train_batch_size:,}" ) logger.info_rank0(f" Gradient Accumulation steps = {self.args.gradient_accumulation_steps:,}") logger.info_rank0(f" Num optimization epochs per batch = {self.finetuning_args.ppo_epochs:,}") logger.info_rank0(f" Total training steps = {max_steps:,}") logger.info_rank0(f" Number of trainable parameters = {count_parameters(self.model)[0]:,}") dataiter = iter(self.dataloader) loss_meter = AverageMeter() reward_meter = AverageMeter() self.callback_handler.on_train_begin(self.args, self.state, self.control) for step in tqdm(range(max_steps), disable=not self.is_local_process_zero()): try: batch = next(dataiter) except StopIteration: dataiter = iter(self.dataloader) batch = next(dataiter) # Get inputs self.model.eval() self.tokenizer.padding_side = "right" # change padding side queries, responses, rewards = [], [], [] for idx in range(0, self.config.batch_size, self.config.mini_batch_size): mini_batch = { "input_ids": batch["input_ids"][idx : idx + self.config.mini_batch_size], "attention_mask": batch["attention_mask"][idx : idx + self.config.mini_batch_size], } mini_batch_queries, mini_batch_responses = self.get_inputs(mini_batch) mini_batch_rewards = self.get_rewards(mini_batch_queries, mini_batch_responses) queries.extend(mini_batch_queries) responses.extend(mini_batch_responses) rewards.extend(mini_batch_rewards) # Run PPO step self.model.train() stats = self.step(queries, responses, rewards) self.tokenizer.padding_side = "left" # restore padding side loss_meter.update(float(stats["ppo/loss/total"]), n=len(rewards)) reward_meter.update(torch.stack(rewards).mean().item(), n=len(rewards)) if self.config.log_with is not None: try: batch["query"] = self.tokenizer.batch_decode(queries, skip_special_tokens=True) batch["response"] = self.tokenizer.batch_decode(responses, skip_special_tokens=True) self.log_stats(stats, batch, rewards) except Exception: logger.warning_rank0("Failed to save stats due to unknown errors.") self.state.global_step += 1 self.callback_handler.on_step_end(self.args, self.state, self.control) if self.is_local_process_zero() and (step + 1) % self.args.logging_steps == 0: logs = dict( loss=round(loss_meter.avg, 4), reward=round(reward_meter.avg, 4), learning_rate=stats["ppo/learning_rate"], epoch=round(step / steps_in_epoch, 2), ) tqdm.write(str(logs)) logs["step"] = step self.state.log_history.append(logs) self.callback_handler.on_log(self.args, self.state, self.control, logs) loss_meter.reset() reward_meter.reset() if (step + 1) % self.args.save_steps == 0: # save checkpoint self.save_model( os.path.join(self.args.output_dir, f"{PREFIX_CHECKPOINT_DIR}-{self.state.global_step}") ) self.callback_handler.on_save(self.args, self.state, self.control) if self.control.should_epoch_stop or self.control.should_training_stop: break self.callback_handler.on_train_end(self.args, self.state, self.control) @override def create_optimizer( self, model: "AutoModelForCausalLMWithValueHead", training_args: "Seq2SeqTrainingArguments", finetuning_args: "FinetuningArguments", ) -> "torch.optim.Optimizer": optimizer = create_custom_optimizer(model, training_args, finetuning_args) if optimizer is None: decay_params, nodecay_params = [], [] decay_param_names = self.get_decay_parameter_names(model) for name, param in model.named_parameters(): if param.requires_grad: if name in decay_param_names: decay_params.append(param) else: nodecay_params.append(param) optim_class, optim_kwargs = Trainer.get_optimizer_cls_and_kwargs(training_args) param_groups = [ dict(params=nodecay_params), dict(params=decay_params, weight_decay=training_args.weight_decay), ] optimizer = optim_class(param_groups, **optim_kwargs) return optimizer @override def create_scheduler( self, training_args: "Seq2SeqTrainingArguments", num_training_steps: int, optimizer: "torch.optim.Optimizer" ) -> "torch.optim.lr_scheduler.LRScheduler": create_custom_scheduler(training_args, num_training_steps, optimizer) lr_scheduler = get_scheduler( training_args.lr_scheduler_type, optimizer=optimizer, num_warmup_steps=training_args.get_warmup_steps(num_training_steps), num_training_steps=num_training_steps, ) return lr_scheduler @torch.no_grad() def get_inputs(self, batch: dict[str, "torch.Tensor"]) -> tuple[list["torch.Tensor"], list["torch.Tensor"]]: r"""Generate model's responses given queries.""" if batch["input_ids"].size(0) == 1: # handle llama2 ppo with gradient accumulation > 1 start_index = (batch["input_ids"][0] != self.tokenizer.pad_token_id).nonzero()[0].item() for k, v in batch.items(): batch[k] = v[:, start_index:] with unwrap_model_for_generation(self.model, self.accelerator) as unwrapped_model: unwrapped_model: AutoModelForCausalLMWithValueHead = self.accelerator.unwrap_model(self.model) if self.model_args.upcast_layernorm: layernorm_params = dump_layernorm(unwrapped_model) generate_output: torch.Tensor = unwrapped_model.generate( generation_config=self.generation_config, logits_processor=get_logits_processor(), **batch ) if self.model_args.upcast_layernorm: restore_layernorm(unwrapped_model, layernorm_params) query = batch["input_ids"].detach().cpu() response = generate_output[:, batch["input_ids"].size(-1) :].detach().cpu() queries, responses = [], [] for i in range(len(query)): query_start_index = (query[i] != self.tokenizer.pad_token_id).nonzero()[0].item() response_indexes = (response[i] != self.tokenizer.pad_token_id).nonzero() if len(response_indexes) == 0: # allow empty response response_length = 1 elif self.tokenizer.eos_token_id == self.tokenizer.pad_token_id: # include eos token response_length = response_indexes[-1].item() + 2 else: response_length = response_indexes[-1].item() + 1 queries.append(query[i, query_start_index:]) # remove padding from left responses.append(response[i, :response_length]) # remove padding from right return queries, responses @torch.no_grad() def get_rewards( self, queries: list["torch.Tensor"], responses: list["torch.Tensor"], ) -> list["torch.Tensor"]: r"""Compute scores using given reward model. Both inputs and outputs are put on CPU. """ if self.finetuning_args.reward_model_type == "api": token_ids = [torch.cat((q, r), dim=-1).tolist() for q, r in zip(queries, responses)] messages = self.tokenizer.batch_decode(token_ids, skip_special_tokens=False) return get_rewards_from_server(self.reward_model, messages) batch: dict[str, torch.Tensor] = self.prepare_model_inputs(queries, responses) unwrapped_model: AutoModelForCausalLMWithValueHead = self.accelerator.unwrap_model(self.model) if self.finetuning_args.reward_model_type in ["lora", "oft"]: replace_model(unwrapped_model, target="reward") reward_model = self.model else: reward_model = self.reward_model with unwrap_model_for_generation(reward_model, self.accelerator), self.amp_context: # support bf16 values: torch.Tensor = reward_model(**batch, return_dict=True, use_cache=False)[-1] if self.finetuning_args.reward_model_type in ["lora", "oft"]: replace_model(unwrapped_model, target="default") rewards = values.gather(dim=-1, index=(batch["attention_mask"].sum(dim=-1, keepdim=True) - 1)) return rewards.float().detach() # use fp32 type @override def batched_forward_pass( self, model: "AutoModelForCausalLMWithValueHead", queries: "torch.Tensor", responses: "torch.Tensor", model_inputs: dict[str, Any], return_logits: bool = False, response_masks: Optional["torch.Tensor"] = None, ) -> tuple["torch.Tensor", Optional["torch.Tensor"], "torch.Tensor", "torch.Tensor"]: r"""Calculate model outputs in multiple batches. Subclass and override to inject custom behavior. """ from trl.core import logprobs_from_logits torch_gc() bs = len(queries) fbs = self.config.mini_batch_size all_logprobs = [] all_logits = [] all_masks = [] all_values = [] for i in range(math.ceil(bs / fbs)): input_kwargs = {key: value[i * fbs : (i + 1) * fbs] for key, value in model_inputs.items()} query_batch = queries[i * fbs : (i + 1) * fbs] response_batch = responses[i * fbs : (i + 1) * fbs] if response_masks is not None: response_masks_batch = response_masks[i * fbs : (i + 1) * fbs] input_ids = input_kwargs["input_ids"] attention_mask = input_kwargs["attention_mask"] with self.amp_context: # support bf16 logits, _, values = model(**input_kwargs, return_dict=True, use_cache=False) logprobs = logprobs_from_logits(logits[:, :-1, :], input_ids[:, 1:]) masks = torch.zeros_like(attention_mask) masks[:, :-1] = attention_mask[:, 1:] for j in range(len(query_batch)): start = len(query_batch[j]) - 1 if attention_mask[j, 0] == 0: # offset left padding start += attention_mask[j, :].nonzero()[0].item() end = start + len(response_batch[j]) if response_masks is not None: response_masks_batch = torch.cat((torch.zeros_like(query_batch[j]), response_masks_batch[j]))[1:] masks[j, :start] = 0 masks[j, end:] = 0 if response_masks is not None: masks[j, start:end] = masks[j, start:end] * response_masks_batch[j][start:end] if return_logits: all_logits.append(logits) else: del logits all_values.append(values) all_logprobs.append(logprobs) all_masks.append(masks) return ( torch.cat(all_logprobs), torch.cat(all_logits)[:, :-1] if return_logits else None, torch.cat(all_values)[:, :-1], torch.cat(all_masks)[:, :-1], ) @override def save_model(self, output_dir: Optional[str] = None) -> None: r"""Save model checkpoint. Subclass and override to inject custom behavior. """ if output_dir is None: output_dir = self.args.output_dir if self.is_fsdp_enabled or self.is_deepspeed_enabled: try: state_dict = self.accelerator.get_state_dict(self.model) # must be called at all ranks if self.args.should_save: self._save(output_dir, state_dict=state_dict) except ValueError: logger.warning_rank0( " stage3_gather_16bit_weights_on_model_save=false. Saving the full checkpoint instead," " use zero_to_fp32.py to recover weights" ) if self.args.should_save: self._save(output_dir, state_dict={}) # remove the dummy state_dict remove_dummy_checkpoint(self.args.should_save, output_dir, [WEIGHTS_NAME, SAFE_WEIGHTS_NAME]) self.model.save_checkpoint(output_dir) elif self.args.should_save: unwrapped_model: AutoModelForCausalLMWithValueHead = self.accelerator.unwrap_model(self.model) self._save(output_dir, state_dict=unwrapped_model.state_dict())
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/train/ppo/__init__.py
src/llamafactory/train/ppo/__init__.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from .workflow import run_ppo __all__ = ["run_ppo"]
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/train/ppo/workflow.py
src/llamafactory/train/ppo/workflow.py
# Copyright 2025 HuggingFace Inc. and the LlamaFactory team. # # This code is inspired by the HuggingFace's TRL library. # https://github.com/huggingface/trl/blob/v0.8.0/examples/scripts/ppo.py # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING, Optional from ...data import MultiModalDataCollatorForSeq2Seq, get_dataset, get_template_and_fix_tokenizer from ...extras.ploting import plot_loss from ...model import load_model, load_tokenizer from ..callbacks import fix_valuehead_checkpoint from ..trainer_utils import create_ref_model, create_reward_model from .trainer import CustomPPOTrainer if TYPE_CHECKING: from transformers import Seq2SeqTrainingArguments, TrainerCallback from ...hparams import DataArguments, FinetuningArguments, GeneratingArguments, ModelArguments def run_ppo( model_args: "ModelArguments", data_args: "DataArguments", training_args: "Seq2SeqTrainingArguments", finetuning_args: "FinetuningArguments", generating_args: "GeneratingArguments", callbacks: Optional[list["TrainerCallback"]] = None, ): tokenizer_module = load_tokenizer(model_args) tokenizer = tokenizer_module["tokenizer"] template = get_template_and_fix_tokenizer(tokenizer, data_args) dataset_module = get_dataset(template, model_args, data_args, training_args, stage="ppo", **tokenizer_module) model = load_model(tokenizer, model_args, finetuning_args, training_args.do_train, add_valuehead=True) tokenizer.padding_side = "left" # use left-padding in generation while using right-padding in training data_collator = MultiModalDataCollatorForSeq2Seq(template=template, model=model, **tokenizer_module) # Create reference model and reward model ref_model = create_ref_model(model_args, finetuning_args, add_valuehead=True) reward_model = create_reward_model(model, model_args, finetuning_args) # Initialize our Trainer ppo_trainer: CustomPPOTrainer = CustomPPOTrainer( model_args=model_args, training_args=training_args, finetuning_args=finetuning_args, generating_args=generating_args, callbacks=callbacks, model=model, reward_model=reward_model, ref_model=ref_model, data_collator=data_collator, **dataset_module, **tokenizer_module, ) # Training if training_args.do_train: ppo_trainer.ppo_train(resume_from_checkpoint=training_args.resume_from_checkpoint) ppo_trainer.save_model() if training_args.should_save: fix_valuehead_checkpoint(model, training_args.output_dir, training_args.save_safetensors) ppo_trainer.save_state() # must be called after save_model to have a folder if ppo_trainer.is_world_process_zero() and finetuning_args.plot_loss: plot_loss(training_args.output_dir, keys=["loss", "reward"])
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/train/ppo/ppo_utils.py
src/llamafactory/train/ppo/ppo_utils.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import json from contextlib import nullcontext from typing import TYPE_CHECKING, Literal, Optional import torch from transformers.integrations import is_deepspeed_zero3_enabled from ...extras.packages import is_requests_available if is_requests_available(): import requests if TYPE_CHECKING: from transformers import PreTrainedModel from trl import AutoModelForCausalLMWithValueHead def get_rewards_from_server(server_url: str, messages: list[str]) -> list["torch.Tensor"]: r"""Get reward scores from the API server.""" headers = {"Content-Type": "application/json"} payload = {"model": "model", "messages": messages} response = requests.post(server_url, json=payload, headers=headers) rewards = json.loads(response.text)["scores"] return torch.Tensor(rewards) def replace_model(model: "AutoModelForCausalLMWithValueHead", target: Literal["default", "reward"]) -> None: r"""Replace the default/reward modules in the model. The model is already unwrapped.""" v_head_layer = model.v_head.summary if is_deepspeed_zero3_enabled(): import deepspeed # type: ignore params = [v_head_layer.weight, v_head_layer.bias] context_maybe_zero3 = deepspeed.zero.GatheredParameters(params, modifier_rank=0) else: context_maybe_zero3 = nullcontext() model.pretrained_model.set_adapter(target) # set the LoRA adapter to be active with context_maybe_zero3: if target == "reward": # save default head temporarily setattr(model, "default_head_weight", v_head_layer.weight.data.detach().clone()) setattr(model, "default_head_bias", v_head_layer.bias.data.detach().clone()) device = v_head_layer.weight.device v_head_layer.weight.data = model.get_buffer(f"{target}_head_weight").detach().clone().to(device) v_head_layer.bias.data = model.get_buffer(f"{target}_head_bias").detach().clone().to(device) def dump_layernorm(model: "PreTrainedModel") -> dict[str, "torch.Tensor"]: r"""Dump the layernorm parameters in the model. The model is already unwrapped (and gathered).""" layer_norm_params = {} for name, param in model.named_parameters(): if param.data.dtype == torch.float32: layer_norm_params[name] = param.data.detach().clone() param.data = param.data.to(model.config.torch_dtype) return layer_norm_params def restore_layernorm(model: "PreTrainedModel", layernorm_params: Optional[dict[str, "torch.Tensor"]] = None) -> None: r"""Restore the layernorm parameters in the model. The model is already unwrapped (and gathered).""" for name, param in model.named_parameters(): if name in layernorm_params: param.data = layernorm_params[name]
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/train/sft/trainer.py
src/llamafactory/train/sft/trainer.py
# Copyright 2025 HuggingFace Inc. and the LlamaFactory team. # # This code is inspired by the HuggingFace's transformers library. # https://github.com/huggingface/transformers/blob/v4.40.0/src/transformers/trainer_seq2seq.py # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import json import os from types import MethodType from typing import TYPE_CHECKING, Any, Optional, Union import numpy as np import torch from transformers import Seq2SeqTrainer from typing_extensions import override from ...extras import logging from ...extras.constants import IGNORE_INDEX from ..callbacks import SaveProcessorCallback from ..fp8_utils import configure_fp8_environment, patch_accelerator_for_fp8, verify_fp8_status from ..trainer_utils import create_custom_optimizer, create_custom_scheduler if TYPE_CHECKING: from torch.utils.data import Dataset from transformers import ProcessorMixin from transformers.trainer import PredictionOutput from ...hparams import FinetuningArguments, ModelArguments, TrainingArguments logger = logging.get_logger(__name__) class CustomSeq2SeqTrainer(Seq2SeqTrainer): r"""Inherits Seq2SeqTrainer to compute generative metrics such as BLEU and ROUGE.""" def __init__( self, finetuning_args: "FinetuningArguments", processor: Optional["ProcessorMixin"], model_args: Optional["ModelArguments"] = None, gen_kwargs: Optional[dict[str, Any]] = None, **kwargs, ) -> None: kwargs["processing_class"] = kwargs.pop("tokenizer") # Configure FP8 environment if enabled training_args: TrainingArguments = kwargs.get("args") if training_args.fp8: configure_fp8_environment(training_args) if getattr(training_args, "fp8_backend", "auto") == "te": patch_accelerator_for_fp8() super().__init__(**kwargs) if processor is not None: # avoid wrong loss under gradient accumulation # https://github.com/huggingface/transformers/pull/36044#issuecomment-2746657112 self.model_accepts_loss_kwargs = False self.finetuning_args = finetuning_args if gen_kwargs is not None: # https://github.com/huggingface/transformers/blob/v4.45.0/src/transformers/trainer_seq2seq.py#L287 self._gen_kwargs = gen_kwargs if processor is not None: self.add_callback(SaveProcessorCallback(processor)) if finetuning_args.use_badam: from badam import BAdamCallback, clip_grad_norm_old_version # type: ignore self.accelerator.clip_grad_norm_ = MethodType(clip_grad_norm_old_version, self.accelerator) self.add_callback(BAdamCallback) if finetuning_args.use_dft_loss: from ..trainer_utils import dft_loss_func self.compute_loss_func = dft_loss_func if training_args.fp8 and hasattr(self, "accelerator"): # verify FP8 status after trainer initialization verify_fp8_status(self.accelerator, training_args) @override def create_optimizer(self) -> "torch.optim.Optimizer": if self.optimizer is None: self.optimizer = create_custom_optimizer(self.model, self.args, self.finetuning_args) return super().create_optimizer() @override def create_scheduler( self, num_training_steps: int, optimizer: Optional["torch.optim.Optimizer"] = None ) -> "torch.optim.lr_scheduler.LRScheduler": create_custom_scheduler(self.args, num_training_steps, optimizer) return super().create_scheduler(num_training_steps, optimizer) @override def _get_train_sampler(self, *args, **kwargs) -> Optional["torch.utils.data.Sampler"]: if self.finetuning_args.disable_shuffling: return torch.utils.data.SequentialSampler(self.train_dataset) return super()._get_train_sampler(*args, **kwargs) @override def compute_loss(self, model, inputs, *args, **kwargs): return super().compute_loss(model, inputs, *args, **kwargs) @override def prediction_step( self, model: "torch.nn.Module", inputs: dict[str, Union["torch.Tensor", Any]], prediction_loss_only: bool, ignore_keys: Optional[list[str]] = None, **gen_kwargs, ) -> tuple[Optional[float], Optional["torch.Tensor"], Optional["torch.Tensor"]]: r"""Remove the prompt part in the generated tokens. Subclass and override to inject custom behavior. """ if self.args.predict_with_generate: # do not pass labels to model when generate labels = inputs.pop("labels", None) else: labels = inputs.get("labels") loss, generated_tokens, _ = super().prediction_step( model, inputs, prediction_loss_only=prediction_loss_only, ignore_keys=ignore_keys, **gen_kwargs ) if generated_tokens is not None and self.args.predict_with_generate: generated_tokens[:, : inputs["input_ids"].size(-1)] = self.processing_class.pad_token_id generated_tokens = generated_tokens.contiguous() return loss, generated_tokens, labels def save_predictions( self, dataset: "Dataset", predict_results: "PredictionOutput", skip_special_tokens: bool = True ) -> None: r"""Save model predictions to `output_dir`. A custom behavior that not contained in Seq2SeqTrainer. """ if not self.is_world_process_zero(): return output_prediction_file = os.path.join(self.args.output_dir, "generated_predictions.jsonl") logger.info_rank0(f"Saving prediction results to {output_prediction_file}") labels = np.where( predict_results.label_ids != IGNORE_INDEX, predict_results.label_ids, self.processing_class.pad_token_id ) preds = np.where( predict_results.predictions != IGNORE_INDEX, predict_results.predictions, self.processing_class.pad_token_id, ) for i in range(len(preds)): pad_len = np.nonzero(preds[i] != self.processing_class.pad_token_id)[0] if len(pad_len): # move pad token to last preds[i] = np.concatenate((preds[i][pad_len[0] :], preds[i][: pad_len[0]]), axis=-1) decoded_inputs = self.processing_class.batch_decode(dataset["input_ids"], skip_special_tokens=False) decoded_preds = self.processing_class.batch_decode(preds, skip_special_tokens=skip_special_tokens) decoded_labels = self.processing_class.batch_decode(labels, skip_special_tokens=skip_special_tokens) with open(output_prediction_file, "w", encoding="utf-8") as f: for text, pred, label in zip(decoded_inputs, decoded_preds, decoded_labels): f.write(json.dumps({"prompt": text, "predict": pred, "label": label}, ensure_ascii=False) + "\n")
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/train/sft/__init__.py
src/llamafactory/train/sft/__init__.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from .workflow import run_sft __all__ = ["run_sft"]
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/train/sft/workflow.py
src/llamafactory/train/sft/workflow.py
# Copyright 2025 HuggingFace Inc. and the LlamaFactory team. # # This code is inspired by the HuggingFace's transformers library. # https://github.com/huggingface/transformers/blob/v4.40.0/examples/pytorch/summarization/run_summarization.py # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING, Optional from ...data import SFTDataCollatorWith4DAttentionMask, get_dataset, get_template_and_fix_tokenizer from ...extras.constants import IGNORE_INDEX from ...extras.logging import get_logger from ...extras.misc import calculate_tps from ...extras.packages import is_transformers_version_greater_than from ...extras.ploting import plot_loss from ...model import load_model, load_tokenizer from ..trainer_utils import create_modelcard_and_push from .metric import ComputeAccuracy, ComputeSimilarity, eval_logit_processor from .trainer import CustomSeq2SeqTrainer if TYPE_CHECKING: from transformers import Seq2SeqTrainingArguments, TrainerCallback from ...hparams import DataArguments, FinetuningArguments, GeneratingArguments, ModelArguments logger = get_logger(__name__) def run_sft( model_args: "ModelArguments", data_args: "DataArguments", training_args: "Seq2SeqTrainingArguments", finetuning_args: "FinetuningArguments", generating_args: "GeneratingArguments", callbacks: Optional[list["TrainerCallback"]] = None, ): tokenizer_module = load_tokenizer(model_args) tokenizer = tokenizer_module["tokenizer"] template = get_template_and_fix_tokenizer(tokenizer, data_args) dataset_module = get_dataset(template, model_args, data_args, training_args, stage="sft", **tokenizer_module) model = load_model(tokenizer, model_args, finetuning_args, training_args.do_train) if getattr(model, "is_quantized", False) and not training_args.do_train: setattr(model, "_hf_peft_config_loaded", True) # hack here: make model compatible with prediction data_collator = SFTDataCollatorWith4DAttentionMask( template=template, model=model if not training_args.predict_with_generate else None, pad_to_multiple_of=8 if training_args.do_train else None, # for shift short attention label_pad_token_id=IGNORE_INDEX if data_args.ignore_pad_token_for_loss else tokenizer.pad_token_id, block_diag_attn=model_args.block_diag_attn, attn_implementation=getattr(model.config, "_attn_implementation", None), compute_dtype=model_args.compute_dtype, **tokenizer_module, ) # Metric utils metric_module = {} if model_args.use_kt: if training_args.predict_with_generate: raise NotImplementedError("`predict_with_generate` is not supported in KTransformers SFT yet.") elif finetuning_args.compute_accuracy: raise NotImplementedError("`compute_accuracy` is not supported in KTransformers SFT yet.") if training_args.predict_with_generate: metric_module["compute_metrics"] = ComputeSimilarity(tokenizer=tokenizer) elif finetuning_args.compute_accuracy: metric_module["compute_metrics"] = ComputeAccuracy() metric_module["preprocess_logits_for_metrics"] = eval_logit_processor # Keyword arguments for `model.generate` gen_kwargs = generating_args.to_dict(obey_generation_config=True) # Compatible with Transformers v4 and Transformers v5 if is_transformers_version_greater_than("4.58.0"): extra_ids = getattr(tokenizer, "additional_special_tokens_ids", None) if not isinstance(extra_ids, list): extra_special_tokens = getattr(tokenizer, "_extra_special_tokens", []) string_tokens = [str(t) for t in extra_special_tokens] extra_ids = tokenizer.convert_tokens_to_ids(string_tokens) all_eos_ids = [tokenizer.eos_token_id] + [i for i in extra_ids if i != -1] unique_eos_ids = list(dict.fromkeys(all_eos_ids)) gen_kwargs["eos_token_id"] = unique_eos_ids else: gen_kwargs["eos_token_id"] = [tokenizer.eos_token_id] + tokenizer.additional_special_tokens_ids gen_kwargs["pad_token_id"] = tokenizer.pad_token_id # Initialize our Trainer if model_args.use_kt: from ktransformers.sft.lora import KTrainer # type: ignore from ktransformers.util.globals import GLOBAL_CONFIG # type: ignore GLOBAL_CONFIG._config["mod"] = "sft" trainer = KTrainer( model=model, args=training_args, tokenizer=tokenizer_module, data_collator=data_collator, callbacks=callbacks, **dataset_module, **metric_module, ) trainer.model_accepts_loss_kwargs = False model.config.use_cache = False else: trainer = CustomSeq2SeqTrainer( model=model, args=training_args, finetuning_args=finetuning_args, data_collator=data_collator, callbacks=callbacks, gen_kwargs=gen_kwargs, **dataset_module, **tokenizer_module, **metric_module, ) # Training if training_args.do_train: train_result = trainer.train(resume_from_checkpoint=training_args.resume_from_checkpoint) trainer.save_model() if finetuning_args.include_effective_tokens_per_second: train_result.metrics["effective_tokens_per_sec"] = calculate_tps( dataset_module["train_dataset"], train_result.metrics, stage="sft" ) trainer.log_metrics("train", train_result.metrics) trainer.save_metrics("train", train_result.metrics) trainer.save_state() if trainer.is_world_process_zero() and finetuning_args.plot_loss: keys = ["loss"] if isinstance(dataset_module.get("eval_dataset"), dict): keys += sum( [[f"eval_{key}_loss", f"eval_{key}_accuracy"] for key in dataset_module["eval_dataset"].keys()], [] ) else: keys += ["eval_loss", "eval_accuracy"] plot_loss(training_args.output_dir, keys=keys) if training_args.predict_with_generate: tokenizer.padding_side = "left" # use left-padding in generation # Evaluation if training_args.do_eval: metrics = trainer.evaluate(metric_key_prefix="eval", **gen_kwargs) trainer.log_metrics("eval", metrics) trainer.save_metrics("eval", metrics) # Predict if training_args.do_predict: logger.warning_rank0_once("Batch generation can be very slow. Consider using `scripts/vllm_infer.py` instead.") predict_results = trainer.predict(dataset_module["eval_dataset"], metric_key_prefix="predict", **gen_kwargs) trainer.log_metrics("predict", predict_results.metrics) trainer.save_metrics("predict", predict_results.metrics) trainer.save_predictions(dataset_module["eval_dataset"], predict_results, generating_args.skip_special_tokens) # Create model card create_modelcard_and_push(trainer, model_args, data_args, training_args, finetuning_args)
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/train/sft/metric.py
src/llamafactory/train/sft/metric.py
# Copyright 2025 HuggingFace Inc., THUDM, and the LlamaFactory team. # # This code is inspired by the HuggingFace's transformers library and the THUDM's ChatGLM implementation. # https://github.com/huggingface/transformers/blob/v4.40.0/examples/pytorch/summarization/run_summarization.py # https://github.com/THUDM/ChatGLM-6B/blob/main/ptuning/main.py # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from dataclasses import dataclass from typing import TYPE_CHECKING, Optional import numpy as np import torch from transformers.utils import is_nltk_available from ...extras.constants import IGNORE_INDEX from ...extras.misc import numpify from ...extras.packages import is_jieba_available, is_rouge_available if TYPE_CHECKING: from transformers import EvalPrediction, PreTrainedTokenizer if is_jieba_available(): import jieba # type: ignore if is_nltk_available(): from nltk.translate.bleu_score import SmoothingFunction, sentence_bleu # type: ignore if is_rouge_available(): from rouge_chinese import Rouge # type: ignore def eval_logit_processor(logits: "torch.Tensor", labels: "torch.Tensor") -> "torch.Tensor": r"""Compute the token with the largest likelihood to reduce memory footprint.""" if isinstance(logits, (list, tuple)): if logits[0].dim() == 3: # (batch_size, seq_len, vocab_size) logits = logits[0] else: # moe models have aux loss logits = logits[1] if logits.dim() != 3: raise ValueError("Cannot process the logits.") return torch.argmax(logits, dim=-1) @dataclass class ComputeAccuracy: r"""Compute accuracy and support `batch_eval_metrics`.""" def _dump(self) -> Optional[dict[str, float]]: result = None if hasattr(self, "score_dict"): result = {k: float(np.mean(v)) for k, v in self.score_dict.items()} self.score_dict = {"accuracy": []} return result def __post_init__(self): self._dump() def __call__(self, eval_preds: "EvalPrediction", compute_result: bool = True) -> Optional[dict[str, float]]: preds, labels = numpify(eval_preds.predictions), numpify(eval_preds.label_ids) for i in range(len(preds)): pred, label = preds[i, :-1], labels[i, 1:] label_mask = label != IGNORE_INDEX self.score_dict["accuracy"].append(np.mean(pred[label_mask] == label[label_mask])) if compute_result: return self._dump() @dataclass class ComputeSimilarity: r"""Compute text similarity scores and support `batch_eval_metrics`. Wraps the tokenizer into metric functions, used in CustomSeq2SeqTrainer. """ tokenizer: "PreTrainedTokenizer" def _dump(self) -> Optional[dict[str, float]]: result = None if hasattr(self, "score_dict"): result = {k: float(np.mean(v)) for k, v in self.score_dict.items()} self.score_dict = {"rouge-1": [], "rouge-2": [], "rouge-l": [], "bleu-4": []} return result def __post_init__(self): self._dump() def __call__(self, eval_preds: "EvalPrediction", compute_result: bool = True) -> Optional[dict[str, float]]: preds, labels = numpify(eval_preds.predictions), numpify(eval_preds.label_ids) preds = np.where(preds != IGNORE_INDEX, preds, self.tokenizer.pad_token_id) labels = np.where(labels != IGNORE_INDEX, labels, self.tokenizer.pad_token_id) decoded_preds = self.tokenizer.batch_decode(preds, skip_special_tokens=True) decoded_labels = self.tokenizer.batch_decode(labels, skip_special_tokens=True) for pred, label in zip(decoded_preds, decoded_labels): hypothesis = list(jieba.cut(pred)) reference = list(jieba.cut(label)) if len(" ".join(hypothesis).split()) == 0 or len(" ".join(reference).split()) == 0: result = {"rouge-1": {"f": 0.0}, "rouge-2": {"f": 0.0}, "rouge-l": {"f": 0.0}} else: rouge = Rouge() scores = rouge.get_scores(" ".join(hypothesis), " ".join(reference)) result = scores[0] for k, v in result.items(): self.score_dict[k].append(round(v["f"] * 100, 4)) bleu_score = sentence_bleu([list(label)], list(pred), smoothing_function=SmoothingFunction().method3) self.score_dict["bleu-4"].append(round(bleu_score * 100, 4)) if compute_result: return self._dump()
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/train/pt/trainer.py
src/llamafactory/train/pt/trainer.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from types import MethodType from typing import TYPE_CHECKING, Optional import torch from transformers import Trainer from typing_extensions import override from ..callbacks import SaveProcessorCallback from ..fp8_utils import configure_fp8_environment, patch_accelerator_for_fp8, verify_fp8_status from ..trainer_utils import create_custom_optimizer, create_custom_scheduler if TYPE_CHECKING: from transformers import ProcessorMixin from ...hparams import FinetuningArguments, ModelArguments, TrainingArguments class CustomTrainer(Trainer): r"""Inherit Trainer for custom optimizer.""" def __init__( self, finetuning_args: "FinetuningArguments", processor: Optional["ProcessorMixin"], model_args: Optional["ModelArguments"] = None, **kwargs, ) -> None: kwargs["processing_class"] = kwargs.pop("tokenizer") # Configure FP8 environment if enabled training_args: TrainingArguments = kwargs.get("args") if training_args.fp8: configure_fp8_environment(training_args) if getattr(training_args, "fp8_backend", "auto") == "te": patch_accelerator_for_fp8() super().__init__(**kwargs) if processor is not None: # avoid wrong loss under gradient accumulation # https://github.com/huggingface/transformers/pull/36044#issuecomment-2746657112 self.model_accepts_loss_kwargs = False self.finetuning_args = finetuning_args if processor is not None: self.add_callback(SaveProcessorCallback(processor)) if finetuning_args.use_badam: from badam import BAdamCallback, clip_grad_norm_old_version # type: ignore self.accelerator.clip_grad_norm_ = MethodType(clip_grad_norm_old_version, self.accelerator) self.add_callback(BAdamCallback) if training_args.fp8 and hasattr(self, "accelerator"): # verify FP8 status after trainer initialization verify_fp8_status(self.accelerator, training_args) @override def create_optimizer(self) -> "torch.optim.Optimizer": if self.optimizer is None: self.optimizer = create_custom_optimizer(self.model, self.args, self.finetuning_args) return super().create_optimizer() @override def create_scheduler( self, num_training_steps: int, optimizer: Optional["torch.optim.Optimizer"] = None ) -> "torch.optim.lr_scheduler.LRScheduler": create_custom_scheduler(self.args, num_training_steps, optimizer) return super().create_scheduler(num_training_steps, optimizer) @override def _get_train_sampler(self, *args, **kwargs) -> Optional["torch.utils.data.Sampler"]: if self.finetuning_args.disable_shuffling: return torch.utils.data.SequentialSampler(self.train_dataset) return super()._get_train_sampler(*args, **kwargs) @override def compute_loss(self, model, inputs, *args, **kwargs): return super().compute_loss(model, inputs, *args, **kwargs)
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/train/pt/__init__.py
src/llamafactory/train/pt/__init__.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from .workflow import run_pt __all__ = ["run_pt"]
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/train/pt/workflow.py
src/llamafactory/train/pt/workflow.py
# Copyright 2025 HuggingFace Inc. and the LlamaFactory team. # # This code is inspired by the HuggingFace's transformers library. # https://github.com/huggingface/transformers/blob/v4.40.0/examples/pytorch/language-modeling/run_clm.py # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import math from typing import TYPE_CHECKING, Optional from transformers import DataCollatorForLanguageModeling from ...data import get_dataset, get_template_and_fix_tokenizer from ...extras.ploting import plot_loss from ...model import load_model, load_tokenizer from ..trainer_utils import create_modelcard_and_push from .trainer import CustomTrainer if TYPE_CHECKING: from transformers import Seq2SeqTrainingArguments, TrainerCallback from ...hparams import DataArguments, FinetuningArguments, ModelArguments def run_pt( model_args: "ModelArguments", data_args: "DataArguments", training_args: "Seq2SeqTrainingArguments", finetuning_args: "FinetuningArguments", callbacks: Optional[list["TrainerCallback"]] = None, ): tokenizer_module = load_tokenizer(model_args) tokenizer = tokenizer_module["tokenizer"] template = get_template_and_fix_tokenizer(tokenizer, data_args) dataset_module = get_dataset(template, model_args, data_args, training_args, stage="pt", **tokenizer_module) model = load_model(tokenizer, model_args, finetuning_args, training_args.do_train) data_collator = DataCollatorForLanguageModeling(tokenizer=tokenizer, mlm=False) # Initialize our Trainer trainer = CustomTrainer( model=model, args=training_args, finetuning_args=finetuning_args, data_collator=data_collator, callbacks=callbacks, **dataset_module, **tokenizer_module, ) # Training if training_args.do_train: train_result = trainer.train(resume_from_checkpoint=training_args.resume_from_checkpoint) trainer.save_model() trainer.log_metrics("train", train_result.metrics) trainer.save_metrics("train", train_result.metrics) trainer.save_state() if trainer.is_world_process_zero() and finetuning_args.plot_loss: keys = ["loss"] if isinstance(dataset_module.get("eval_dataset"), dict): keys += [f"eval_{key}_loss" for key in dataset_module["eval_dataset"].keys()] else: keys += ["eval_loss"] plot_loss(training_args.output_dir, keys=keys) # Evaluation if training_args.do_eval: metrics = trainer.evaluate(metric_key_prefix="eval") if isinstance(dataset_module.get("eval_dataset"), dict): for key in dataset_module["eval_dataset"].keys(): try: perplexity = math.exp(metrics[f"eval_{key}_loss"]) except OverflowError: perplexity = float("inf") metrics[f"eval_{key}_perplexity"] = perplexity else: try: perplexity = math.exp(metrics["eval_loss"]) except OverflowError: perplexity = float("inf") metrics["eval_perplexity"] = perplexity trainer.log_metrics("eval", metrics) trainer.save_metrics("eval", metrics) # Create model card create_modelcard_and_push(trainer, model_args, data_args, training_args, finetuning_args)
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/train/dpo/trainer.py
src/llamafactory/train/dpo/trainer.py
# Copyright 2025 HuggingFace Inc. and the LlamaFactory team. # # This code is inspired by the HuggingFace's TRL library. # https://github.com/huggingface/trl/blob/v0.8.0/trl/trainer/dpo_trainer.py # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import warnings from collections import defaultdict from contextlib import nullcontext from types import MethodType from typing import TYPE_CHECKING, Literal, Optional, Union import torch import torch.nn.functional as F from transformers import Trainer from trl import DPOTrainer from trl.trainer import disable_dropout_in_model from trl.trainer.utils import prepare_deepspeed from typing_extensions import override from ...extras.constants import IGNORE_INDEX from ...extras.packages import is_transformers_version_greater_than from ..callbacks import SaveProcessorCallback from ..trainer_utils import create_custom_optimizer, create_custom_scheduler, get_batch_logps, nested_detach if TYPE_CHECKING: from transformers import PreTrainedModel, ProcessorMixin from ...hparams import FinetuningArguments class CustomDPOTrainer(DPOTrainer): def __init__( self, model: Union["PreTrainedModel", torch.nn.Module], ref_model: Optional[Union["PreTrainedModel", torch.nn.Module]], finetuning_args: "FinetuningArguments", processor: Optional["ProcessorMixin"], disable_dropout: bool = True, **kwargs, ): if is_transformers_version_greater_than("4.46"): kwargs["processing_class"] = kwargs.pop("tokenizer") if disable_dropout: disable_dropout_in_model(model) if ref_model is not None: disable_dropout_in_model(ref_model) self.finetuning_args = finetuning_args self.f_divergence_type = "reverse_kl" self.reference_free = False self.use_dpo_data_collator = True # hack to avoid warning self.generate_during_eval = False # disable at evaluation self.label_pad_token_id = IGNORE_INDEX self.padding_value = 0 self.is_encoder_decoder = model.config.is_encoder_decoder self.precompute_ref_log_probs = False self._precomputed_train_ref_log_probs = False self._precomputed_eval_ref_log_probs = False self._peft_has_been_casted_to_bf16 = False self.ref_model = ref_model self._stored_metrics = defaultdict(lambda: defaultdict(list)) # dpo hyperparams self.beta = finetuning_args.pref_beta self.loss_type = finetuning_args.pref_loss self.ftx_gamma = finetuning_args.pref_ftx self.bco_gemma = finetuning_args.pref_bco_weight self.label_smoothing = finetuning_args.dpo_label_smoothing self.simpo_gamma = finetuning_args.simpo_gamma self.ld_alpha = finetuning_args.ld_alpha Trainer.__init__(self, model=model, **kwargs) self.model_accepts_loss_kwargs = False # overwrite trainer's default behavior if not hasattr(self, "accelerator"): raise AttributeError("Please update `transformers`.") warnings.simplefilter("ignore") # remove gc warnings on ref model if ref_model is not None: if self.is_deepspeed_enabled: if not ( getattr(ref_model, "is_loaded_in_8bit", False) or getattr(ref_model, "is_loaded_in_4bit", False) ): # quantized models are already set on the correct device self.ref_model = prepare_deepspeed(self.ref_model, self.accelerator) else: self.ref_model = self.accelerator.prepare_model(self.ref_model, evaluation_mode=True) self.ref_model.eval() if processor is not None: self.add_callback(SaveProcessorCallback(processor)) if finetuning_args.use_badam: from badam import BAdamCallback, clip_grad_norm_old_version # type: ignore self.accelerator.clip_grad_norm_ = MethodType(clip_grad_norm_old_version, self.accelerator) self.add_callback(BAdamCallback) if self.bco_gemma >= 1e-6: from trl.trainer import RunningMoments self.running = RunningMoments(self.accelerator) @override def create_optimizer(self) -> "torch.optim.Optimizer": if self.optimizer is None: self.optimizer = create_custom_optimizer(self.model, self.args, self.finetuning_args) return super().create_optimizer() @override def create_scheduler( self, num_training_steps: int, optimizer: Optional["torch.optim.Optimizer"] = None ) -> "torch.optim.lr_scheduler.LRScheduler": create_custom_scheduler(self.args, num_training_steps, optimizer) return super().create_scheduler(num_training_steps, optimizer) @override def _get_train_sampler(self, *args, **kwargs) -> Optional["torch.utils.data.Sampler"]: if self.finetuning_args.disable_shuffling: return torch.utils.data.SequentialSampler(self.train_dataset) return super()._get_train_sampler(*args, **kwargs) @override def get_batch_samples(self, *args, **kwargs): r"""Replace the method of DPO Trainer with the one of the standard Trainer.""" return Trainer.get_batch_samples(self, *args, **kwargs) def odds_ratio_loss(self, chosen_logps: "torch.Tensor", rejected_logps: "torch.Tensor") -> "torch.Tensor": r"""Compute ORPO's odds ratio (OR) loss for batched log probabilities of the policy model.""" log_odds = (chosen_logps - rejected_logps) - ( torch.log1p(-torch.exp(chosen_logps)) - torch.log1p(-torch.exp(rejected_logps)) ) sft_loss = -chosen_logps odds_ratio_loss = -F.logsigmoid(log_odds) orpo_loss = sft_loss + self.beta * odds_ratio_loss return orpo_loss def simpo_loss(self, chosen_logps: "torch.Tensor", rejected_logps: "torch.Tensor") -> "torch.Tensor": r"""Compute SimPO loss for batched log probabilities of the policy model.""" pi_logratios = chosen_logps - rejected_logps gamma_logratios = self.simpo_gamma / self.beta logits = pi_logratios - gamma_logratios simpo_loss = -F.logsigmoid(self.beta * logits) return simpo_loss def bco_loss( self, chosen_logps: "torch.Tensor", rejected_logps: "torch.Tensor", reference_chosen_logps: "torch.Tensor", reference_rejected_logps: "torch.Tensor", ) -> "torch.Tensor": chosen_logratios = chosen_logps - reference_chosen_logps rejected_logratios = rejected_logps - reference_rejected_logps chosen_rewards = self.beta * chosen_logratios rejected_rewards = self.beta * rejected_logratios rewards = torch.cat((chosen_rewards, rejected_rewards), 0).mean().detach() self.running.update(rewards) # update baseline delta = self.running.mean bco_loss = -F.logsigmoid((self.beta * chosen_logratios) - delta) - F.logsigmoid( -(self.beta * rejected_logratios - delta) ) return bco_loss def compute_preference_loss( self, policy_chosen_logps: "torch.Tensor", policy_rejected_logps: "torch.Tensor", reference_chosen_logps: Optional["torch.Tensor"], reference_rejected_logps: Optional["torch.Tensor"], ) -> tuple["torch.Tensor", "torch.Tensor", "torch.Tensor"]: r"""Compute loss for preference learning.""" if not self.finetuning_args.use_ref_model: if self.loss_type == "orpo": losses = self.odds_ratio_loss(policy_chosen_logps, policy_rejected_logps) elif self.loss_type == "simpo": losses = self.simpo_loss(policy_chosen_logps, policy_rejected_logps) else: raise NotImplementedError(f"Unknown loss type: {self.loss_type}.") chosen_rewards = self.beta * policy_chosen_logps.to(self.accelerator.device).detach() rejected_rewards = self.beta * policy_rejected_logps.to(self.accelerator.device).detach() else: losses, chosen_rewards, rejected_rewards = self.dpo_loss( policy_chosen_logps, policy_rejected_logps, reference_chosen_logps, reference_rejected_logps ) if self.bco_gemma > 1e-6: bco_losses = self.bco_loss( policy_chosen_logps, policy_rejected_logps, reference_chosen_logps, reference_rejected_logps ) losses = (losses + bco_losses * self.bco_gemma) / (1.0 + self.bco_gemma) # re-weight W_p and W_q return losses, chosen_rewards, rejected_rewards @override def concatenated_forward( self, model: "PreTrainedModel", batch: dict[str, "torch.Tensor"], is_ref_model: bool = False ) -> dict[str, "torch.Tensor"]: r"""Compute the sum log probabilities of the labels under given logits if loss_type is not IPO, ORPO or SimPO. Otherwise the average log probabilities. """ if self.finetuning_args.use_ref_model: batch = nested_detach(batch, clone=True) # avoid error labels = batch.pop("labels") # dpo do not need compute loss in forward all_logits: torch.Tensor = model(**batch, return_dict=True, use_cache=False).logits.to(torch.float32) all_logps, valid_length = get_batch_logps( logits=all_logits, labels=labels, ld_alpha=(self.ld_alpha if not is_ref_model else None) ) if self.loss_type in ["ipo", "orpo", "simpo"]: all_logps = all_logps / valid_length batch_size = batch["input_ids"].size(0) // 2 chosen_logps, rejected_logps = all_logps.split(batch_size, dim=0) chosen_logits, rejected_logits = all_logits.split(batch_size, dim=0) chosen_length, _ = valid_length.split(batch_size, dim=0) if self.loss_type in ["ipo", "orpo", "simpo"]: chosen_logps_avg = chosen_logps else: chosen_logps_avg = chosen_logps / chosen_length return { "chosen_logps": chosen_logps, "rejected_logps": rejected_logps, "chosen_logits": chosen_logits, "rejected_logits": rejected_logits, "chosen_logps_avg": chosen_logps_avg, } @override def compute_reference_log_probs( self, model: "PreTrainedModel", batch: dict[str, "torch.Tensor"] ) -> tuple[Optional["torch.Tensor"], Optional["torch.Tensor"]]: r"""Compute log probabilities of the reference model.""" if not self.finetuning_args.use_ref_model: return None, None if self.ref_model is None: ref_model = model ref_context = self.accelerator.unwrap_model(model).disable_adapter() else: ref_model = self.ref_model ref_context = nullcontext() with torch.no_grad(), ref_context: ref_output = self.concatenated_forward(ref_model, batch, is_ref_model=True) reference_chosen_logps = ref_output["chosen_logps"] reference_rejected_logps = ref_output["rejected_logps"] return reference_chosen_logps, reference_rejected_logps @override def get_batch_loss_metrics( self, model: "PreTrainedModel", batch: dict[str, "torch.Tensor"], train_eval: Literal["train", "eval"] = "train", ) -> tuple["torch.Tensor", dict[str, "torch.Tensor"]]: r"""Compute the DPO loss and other metrics for the given batch of inputs for train or test.""" metrics = {} model_output = self.concatenated_forward(model, batch) policy_chosen_logps = model_output["chosen_logps"] policy_rejected_logps = model_output["rejected_logps"] policy_chosen_logits = model_output["chosen_logits"] policy_rejected_logits = model_output["rejected_logits"] policy_chosen_logps_avg = model_output["chosen_logps_avg"] reference_chosen_logps, reference_rejected_logps = self.compute_reference_log_probs(model, batch) losses, chosen_rewards, rejected_rewards = self.compute_preference_loss( policy_chosen_logps, policy_rejected_logps, reference_chosen_logps, reference_rejected_logps, ) sft_loss = -policy_chosen_logps_avg if self.ftx_gamma > 1e-6: losses += self.ftx_gamma * sft_loss prefix = "eval_" if train_eval == "eval" else "" metrics[f"{prefix}rewards/chosen"] = chosen_rewards.mean().item() metrics[f"{prefix}rewards/rejected"] = rejected_rewards.mean().item() metrics[f"{prefix}rewards/accuracies"] = (chosen_rewards > rejected_rewards).float().mean().item() metrics[f"{prefix}rewards/margins"] = (chosen_rewards - rejected_rewards).mean().item() metrics[f"{prefix}logps/chosen"] = policy_chosen_logps.mean().item() metrics[f"{prefix}logps/rejected"] = policy_rejected_logps.mean().item() metrics[f"{prefix}logits/chosen"] = policy_chosen_logits.mean().item() metrics[f"{prefix}logits/rejected"] = policy_rejected_logits.mean().item() if self.loss_type == "orpo": metrics[f"{prefix}sft_loss"] = sft_loss.mean().item() metrics[f"{prefix}odds_ratio_loss"] = ((losses - sft_loss) / self.beta).mean().item() return losses.mean(), metrics @override def compute_loss( self, model: "PreTrainedModel", inputs: dict[str, "torch.Tensor"], return_outputs: bool = False, **kwargs ) -> Union["torch.Tensor", tuple["torch.Tensor", list["torch.Tensor"]]]: r"""Subclass and override to accept extra kwargs.""" return super().compute_loss(model, inputs, return_outputs) @override def log(self, logs: dict[str, float], *args, **kwargs) -> None: r"""Log `logs` on the various objects watching training, including stored metrics.""" # logs either has "loss" or "eval_loss" train_eval = "train" if "loss" in logs else "eval" # Add averaged stored metrics to logs key_list, metric_list = [], [] for key, metrics in self._stored_metrics[train_eval].items(): key_list.append(key) metric_list.append(torch.tensor(metrics, dtype=torch.float).to(self.accelerator.device).mean().item()) del self._stored_metrics[train_eval] if len(metric_list) < 10: # pad to for all reduce for i in range(10 - len(metric_list)): key_list.append(f"dummy_{i}") metric_list.append(0.0) metric_list = torch.tensor(metric_list, dtype=torch.float).to(self.accelerator.device) metric_list = self.accelerator.reduce(metric_list, "mean").tolist() for key, metric in zip(key_list, metric_list): # add remaining items if not key.startswith("dummy_"): logs[key] = metric return Trainer.log(self, logs, *args, **kwargs)
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/train/dpo/__init__.py
src/llamafactory/train/dpo/__init__.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from .workflow import run_dpo __all__ = ["run_dpo"]
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/train/dpo/workflow.py
src/llamafactory/train/dpo/workflow.py
# Copyright 2025 HuggingFace Inc. and the LlamaFactory team. # # This code is inspired by the HuggingFace's TRL library. # https://github.com/huggingface/trl/blob/v0.8.0/examples/scripts/dpo.py # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING, Optional from ...data import PairwiseDataCollatorWithPadding, get_dataset, get_template_and_fix_tokenizer from ...extras.constants import IGNORE_INDEX from ...extras.misc import calculate_tps from ...extras.ploting import plot_loss from ...hparams import ModelArguments from ...model import load_model, load_tokenizer from ..trainer_utils import create_modelcard_and_push, create_ref_model if TYPE_CHECKING: from transformers import Seq2SeqTrainingArguments, TrainerCallback from ...hparams import DataArguments, FinetuningArguments def run_dpo( model_args: "ModelArguments", data_args: "DataArguments", training_args: "Seq2SeqTrainingArguments", finetuning_args: "FinetuningArguments", callbacks: Optional[list["TrainerCallback"]] = None, ): tokenizer_module = load_tokenizer(model_args) tokenizer = tokenizer_module["tokenizer"] template = get_template_and_fix_tokenizer(tokenizer, data_args) dataset_module = get_dataset(template, model_args, data_args, training_args, stage="rm", **tokenizer_module) model = load_model(tokenizer, model_args, finetuning_args, training_args.do_train) data_collator = PairwiseDataCollatorWithPadding( template=template, model=model, pad_to_multiple_of=8, label_pad_token_id=IGNORE_INDEX if data_args.ignore_pad_token_for_loss else tokenizer.pad_token_id, **tokenizer_module, ) # Create reference model if finetuning_args.use_ref_model: if finetuning_args.ref_model is None and (not training_args.do_train): # use the model itself ref_model = model else: ref_model = create_ref_model(model_args, finetuning_args) else: ref_model = None if model_args.use_kt: from ktransformers.util.globals import GLOBAL_CONFIG # type: ignore from .ktrainer import KDPOTrainer as CustomDPOTrainer GLOBAL_CONFIG._config["mod"] = "sft" else: from .trainer import CustomDPOTrainer # Initialize our Trainer trainer = CustomDPOTrainer( model=model, ref_model=ref_model, args=training_args, finetuning_args=finetuning_args, data_collator=data_collator, callbacks=callbacks, **dataset_module, **tokenizer_module, ) # Training if training_args.do_train: train_result = trainer.train(resume_from_checkpoint=training_args.resume_from_checkpoint) trainer.save_model() if finetuning_args.include_effective_tokens_per_second: train_result.metrics["effective_tokens_per_sec"] = calculate_tps( dataset_module["train_dataset"], train_result.metrics, stage="rm" ) trainer.log_metrics("train", train_result.metrics) trainer.save_metrics("train", train_result.metrics) trainer.save_state() if trainer.is_world_process_zero() and finetuning_args.plot_loss: keys = ["loss", "rewards/accuracies"] if isinstance(dataset_module.get("eval_dataset"), dict): keys += [f"eval_{key}_loss" for key in dataset_module["eval_dataset"].keys()] else: keys += ["eval_loss"] plot_loss(training_args.output_dir, keys=keys) # Evaluation if training_args.do_eval: metrics = trainer.evaluate(metric_key_prefix="eval") if id(model) == id(ref_model): # unable to compute rewards if reference model is the model itself remove_keys = [key for key in metrics.keys() if "rewards" in key] for key in remove_keys: metrics.pop(key) trainer.log_metrics("eval", metrics) trainer.save_metrics("eval", metrics) # Create model card create_modelcard_and_push(trainer, model_args, data_args, training_args, finetuning_args)
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/train/dpo/ktrainer.py
src/llamafactory/train/dpo/ktrainer.py
# Copyright 2025 HuggingFace Inc. and the LlamaFactory team. # # This code is inspired by the HuggingFace's TRL library. # https://github.com/huggingface/trl/blob/v0.8.0/trl/trainer/dpo_trainer.py # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING import torch from ktransformers.sft.lora import KTrainer # type: ignore from typing_extensions import override from ..trainer_utils import get_batch_logps, nested_detach from .trainer import CustomDPOTrainer if TYPE_CHECKING: from transformers import PreTrainedModel class KDPOTrainer(KTrainer, CustomDPOTrainer): @override def concatenated_forward( self, model: "PreTrainedModel", batch: dict[str, "torch.Tensor"], is_ref_model: bool = False ) -> tuple["torch.Tensor", "torch.Tensor", "torch.Tensor", "torch.Tensor", "torch.Tensor"]: r"""Compute the sum log probabilities of the labels under given logits if loss_type is not IPO, ORPO or SimPO. Otherwise the average log probabilities. """ if self.finetuning_args.use_ref_model: batch = nested_detach(batch, clone=True) # avoid error labels = batch.pop("labels") # dpo do not need compute loss in forward all_logits: torch.Tensor = model(**batch, return_dict=True, use_cache=False).logits.to(torch.float32) all_logits = all_logits.to("cpu") labels = labels.to(all_logits.device) all_logps, valid_length = get_batch_logps( logits=all_logits, labels=labels, ld_alpha=(self.ld_alpha if not is_ref_model else None) ) if self.loss_type in ["ipo", "orpo", "simpo"]: all_logps = all_logps / valid_length batch_size = batch["input_ids"].size(0) // 2 chosen_logps, rejected_logps = all_logps.split(batch_size, dim=0) chosen_logits, rejected_logits = all_logits.split(batch_size, dim=0) chosen_length, _ = valid_length.split(batch_size, dim=0) if self.loss_type in ["ipo", "orpo", "simpo"]: return chosen_logps, rejected_logps, chosen_logits, rejected_logits, chosen_logps else: return chosen_logps, rejected_logps, chosen_logits, rejected_logits, chosen_logps / chosen_length
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/train/kto/trainer.py
src/llamafactory/train/kto/trainer.py
# Copyright 2025 HuggingFace Inc. and the LlamaFactory team. # # This code is inspired by the HuggingFace's TRL library. # https://github.com/huggingface/trl/blob/v0.8.0/trl/trainer/kto_trainer.py # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import warnings from collections import defaultdict from contextlib import nullcontext from types import MethodType from typing import TYPE_CHECKING, Literal, Optional, Union import torch from transformers import Trainer from trl import KTOTrainer from trl.trainer import disable_dropout_in_model from trl.trainer.utils import prepare_deepspeed from typing_extensions import override from ...extras.constants import IGNORE_INDEX from ...extras.packages import is_transformers_version_greater_than from ..callbacks import SaveProcessorCallback from ..trainer_utils import create_custom_optimizer, create_custom_scheduler, get_batch_logps, nested_detach if TYPE_CHECKING: from transformers import PreTrainedModel, ProcessorMixin from ...hparams import FinetuningArguments class CustomKTOTrainer(KTOTrainer): def __init__( self, model: Union["PreTrainedModel", torch.nn.Module], ref_model: Optional[Union["PreTrainedModel", torch.nn.Module]], finetuning_args: "FinetuningArguments", processor: Optional["ProcessorMixin"], disable_dropout: bool = True, **kwargs, ): if is_transformers_version_greater_than("4.46"): kwargs["processing_class"] = kwargs.pop("tokenizer") if disable_dropout: disable_dropout_in_model(model) if ref_model is not None: disable_dropout_in_model(ref_model) self.finetuning_args = finetuning_args self.reference_free = False self.use_dpo_data_collator = True # hack to avoid warning self.generate_during_eval = False # disable at evaluation self.label_pad_token_id = IGNORE_INDEX self.padding_value = 0 self.is_encoder_decoder = model.config.is_encoder_decoder self.precompute_ref_log_probs = False self._precomputed_train_ref_log_probs = False self._precomputed_eval_ref_log_probs = False self._peft_has_been_casted_to_bf16 = False self.ref_model = ref_model self._stored_metrics = defaultdict(lambda: defaultdict(list)) # kto hyperparams self.beta = finetuning_args.pref_beta self.desirable_weight = finetuning_args.kto_chosen_weight self.undesirable_weight = finetuning_args.kto_rejected_weight self.ftx_gamma = finetuning_args.pref_ftx # trl # Not all losses require a KL calculation self.calculate_KL = True if hasattr(self, "loss_type") and self.loss_type in ["apo_zero_unpaired"]: self.calculate_KL = False else: self.loss_type = "kto" Trainer.__init__(self, model=model, **kwargs) self.model_accepts_loss_kwargs = False # overwrite trainer's default behavior if not hasattr(self, "accelerator"): raise AttributeError("Please update `transformers`.") warnings.simplefilter("ignore") # remove gc warnings on ref model if ref_model is not None: if self.is_deepspeed_enabled: if not ( getattr(ref_model, "is_loaded_in_8bit", False) or getattr(ref_model, "is_loaded_in_4bit", False) ): # quantized models are already set on the correct device self.ref_model = prepare_deepspeed(self.ref_model, self.accelerator) else: self.ref_model = self.accelerator.prepare_model(self.ref_model, evaluation_mode=True) self.ref_model.eval() if processor is not None: self.add_callback(SaveProcessorCallback(processor)) if finetuning_args.use_badam: from badam import BAdamCallback, clip_grad_norm_old_version # type: ignore self.accelerator.clip_grad_norm_ = MethodType(clip_grad_norm_old_version, self.accelerator) self.add_callback(BAdamCallback) @override def create_optimizer(self) -> "torch.optim.Optimizer": if self.optimizer is None: self.optimizer = create_custom_optimizer(self.model, self.args, self.finetuning_args) return super().create_optimizer() @override def create_scheduler( self, num_training_steps: int, optimizer: Optional["torch.optim.Optimizer"] = None ) -> "torch.optim.lr_scheduler.LRScheduler": create_custom_scheduler(self.args, num_training_steps, optimizer) return super().create_scheduler(num_training_steps, optimizer) @override def _get_train_sampler(self, *args, **kwargs) -> Optional["torch.utils.data.Sampler"]: r"""Replace the sequential sampler of KTO Trainer created by trl with the random sampler.""" if self.finetuning_args.disable_shuffling: return torch.utils.data.SequentialSampler(self.train_dataset) return Trainer._get_train_sampler(self, *args, **kwargs) @override def get_batch_samples(self, *args, **kwargs): r"""Replace the method of KTO Trainer with the one of the standard Trainer.""" return Trainer.get_batch_samples(self, *args, **kwargs) @override def forward( self, model: "PreTrainedModel", batch: dict[str, "torch.Tensor"], prefix: Literal["", "kl_"] = "" ) -> tuple["torch.Tensor", "torch.Tensor", "torch.Tensor"]: r"""Run forward pass and computes the log probabilities.""" batch = nested_detach(batch, clone=True) # avoid error model_inputs = { "input_ids": batch[f"{prefix}input_ids"], "attention_mask": batch[f"{prefix}attention_mask"], } if f"{prefix}token_type_ids" in batch: model_inputs["token_type_ids"] = batch[f"{prefix}token_type_ids"] if "pixel_values" in batch: model_inputs["pixel_values"] = batch["pixel_values"] if "image_sizes" in batch: model_inputs["image_sizes"] = batch["image_sizes"] if "image_grid_thw" in batch: model_inputs["image_grid_thw"] = batch["image_grid_thw"] if "aspect_ratio_ids" in batch: model_inputs["aspect_ratio_ids"] = batch["aspect_ratio_ids"] if "aspect_ratio_mask" in batch: model_inputs["aspect_ratio_mask"] = batch["aspect_ratio_mask"] if f"{prefix}cross_attention_mask" in batch: model_inputs["cross_attention_mask"] = batch[f"{prefix}cross_attention_mask"] logits = model(**model_inputs, return_dict=True, use_cache=False).logits.to(torch.float32) logps, valid_length = get_batch_logps(logits=logits, labels=batch[f"{prefix}labels"]) return logits, logps, logps / valid_length @override def concatenated_forward( self, model: "PreTrainedModel", batch: dict[str, "torch.Tensor"] ) -> tuple["torch.Tensor", "torch.Tensor", "torch.Tensor", "torch.Tensor", "torch.Tensor", "torch.Tensor"]: target_logits, target_logps, target_logps_avg = self.forward(model, batch) with torch.no_grad(): _, kl_logps, _ = self.forward(model, batch, prefix="kl_") if len(target_logps) != len(batch["kto_tags"]): raise ValueError("Mismatched shape of inputs and labels.") chosen_logits = target_logits[batch["kto_tags"]] chosen_logps = target_logps[batch["kto_tags"]] rejected_logits = target_logits[~batch["kto_tags"]] rejected_logps = target_logps[~batch["kto_tags"]] chosen_logps_avg = target_logps_avg[batch["kto_tags"]] return chosen_logps, rejected_logps, chosen_logits, rejected_logits, kl_logps, chosen_logps_avg @override def compute_reference_log_probs( self, model: "PreTrainedModel", batch: dict[str, "torch.Tensor"] ) -> tuple["torch.Tensor", "torch.Tensor", "torch.Tensor"]: r"""Compute log probabilities of the reference model.""" if self.ref_model is None: ref_model = model ref_context = self.accelerator.unwrap_model(model).disable_adapter() else: ref_model = self.ref_model ref_context = nullcontext() with torch.no_grad(), ref_context: reference_chosen_logps, reference_rejected_logps, _, _, reference_kl_logps, _ = self.concatenated_forward( ref_model, batch ) return reference_chosen_logps, reference_rejected_logps, reference_kl_logps @override def get_batch_loss_metrics( self, model: "PreTrainedModel", batch: dict[str, "torch.Tensor"], ) -> tuple["torch.Tensor", dict[str, "torch.Tensor"]]: r"""Compute the DPO loss and other metrics for the given batch of inputs for train or test.""" metrics = {} ( policy_chosen_logps, policy_rejected_logps, policy_chosen_logits, policy_rejected_logits, policy_kl_logps, policy_chosen_logps_avg, ) = self.concatenated_forward(model, batch) reference_chosen_logps, reference_rejected_logps, reference_kl_logps = self.compute_reference_log_probs( model, batch ) losses, chosen_rewards, rejected_rewards, kl = self.kto_loss( policy_chosen_logps, policy_rejected_logps, policy_kl_logps, reference_chosen_logps, reference_rejected_logps, reference_kl_logps, ) losses = losses.nanmean() if self.ftx_gamma > 1e-6 and len(policy_chosen_logps) > 0: # remember to rescale sft_loss = -policy_chosen_logps_avg losses += self.ftx_gamma * sft_loss.nanmean() / len(policy_chosen_logps) * len(batch["labels"]) num_chosen = len(chosen_rewards) num_rejected = len(rejected_rewards) if num_chosen > 0: metrics["rewards/chosen_sum"] = chosen_rewards.nansum().item() metrics["logps/chosen_sum"] = policy_chosen_logps.nansum().item() metrics["logits/chosen_sum"] = policy_chosen_logits.nansum().item() metrics["count/chosen"] = float(num_chosen) if num_rejected > 0: metrics["rewards/rejected_sum"] = rejected_rewards.nansum().item() metrics["logps/rejected_sum"] = policy_rejected_logps.nansum().item() metrics["logits/rejected_sum"] = policy_rejected_logits.nansum().item() metrics["count/rejected"] = float(num_rejected) metrics["kl"] = kl.item() return losses, metrics @override def compute_loss( self, model: "PreTrainedModel", inputs: dict[str, "torch.Tensor"], return_outputs: bool = False, **kwargs ) -> Union["torch.Tensor", tuple["torch.Tensor", list["torch.Tensor"]]]: r"""Subclass and override to accept extra kwargs.""" return super().compute_loss(model, inputs, return_outputs) @override def log(self, logs: dict[str, float], *args, **kwargs) -> None: r"""Log `logs` on the various objects watching training, including stored metrics.""" # logs either has "loss" or "eval_loss" train_eval = "train" if "loss" in logs else "eval" prefix = "eval_" if train_eval == "eval" else "" # Add averaged stored metrics to logs key_list, metric_list = [], [] for key, metrics in self._stored_metrics[train_eval].items(): key_list.append(key) metric_list.append(torch.tensor(metrics, dtype=torch.float).to(self.accelerator.device).sum().item()) del self._stored_metrics[train_eval] if len(metric_list) < 9: # pad to for all reduce for i in range(9 - len(metric_list)): key_list.append(f"dummy_{i}") metric_list.append(0.0) metric_list = torch.tensor(metric_list, dtype=torch.float).to(self.accelerator.device) metric_list = self.accelerator.reduce(metric_list, "sum").tolist() metric_dict: dict[str, float] = dict(zip(key_list, metric_list)) for split in ["chosen", "rejected"]: # accumulate average metrics from sums and lengths if f"count/{split}" in metric_dict: for key in ("rewards", "logps", "logits"): logs[f"{prefix}{key}/{split}"] = metric_dict[f"{key}/{split}_sum"] / metric_dict[f"count/{split}"] del metric_dict[f"{key}/{split}_sum"] del metric_dict[f"count/{split}"] if f"{prefix}rewards/chosen" in logs and f"{prefix}rewards/rejected" in logs: # calculate reward margin logs[f"{prefix}rewards/margins"] = logs[f"{prefix}rewards/chosen"] - logs[f"{prefix}rewards/rejected"] for key, metric in metric_dict.items(): # add remaining items if not key.startswith("dummy_"): logs[key] = metric return Trainer.log(self, logs, *args, **kwargs)
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/train/kto/__init__.py
src/llamafactory/train/kto/__init__.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from .workflow import run_kto __all__ = ["run_kto"]
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/train/kto/workflow.py
src/llamafactory/train/kto/workflow.py
# Copyright 2025 HuggingFace Inc. and the LlamaFactory team. # # This code is inspired by the HuggingFace's TRL library. # https://github.com/huggingface/trl/blob/v0.8.0/examples/scripts/kto.py # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING, Optional from ...data import KTODataCollatorWithPadding, get_dataset, get_template_and_fix_tokenizer from ...extras.constants import IGNORE_INDEX from ...extras.ploting import plot_loss from ...hparams import ModelArguments from ...model import load_model, load_tokenizer from ..trainer_utils import create_modelcard_and_push, create_ref_model from .trainer import CustomKTOTrainer if TYPE_CHECKING: from transformers import Seq2SeqTrainingArguments, TrainerCallback from ...hparams import DataArguments, FinetuningArguments def run_kto( model_args: "ModelArguments", data_args: "DataArguments", training_args: "Seq2SeqTrainingArguments", finetuning_args: "FinetuningArguments", callbacks: Optional[list["TrainerCallback"]] = None, ): tokenizer_module = load_tokenizer(model_args) tokenizer = tokenizer_module["tokenizer"] template = get_template_and_fix_tokenizer(tokenizer, data_args) dataset_module = get_dataset(template, model_args, data_args, training_args, stage="kto", **tokenizer_module) model = load_model(tokenizer, model_args, finetuning_args, training_args.do_train) data_collator = KTODataCollatorWithPadding( template=template, model=model, pad_to_multiple_of=8, label_pad_token_id=IGNORE_INDEX if data_args.ignore_pad_token_for_loss else tokenizer.pad_token_id, **tokenizer_module, ) # Create reference model if finetuning_args.ref_model is None and (not training_args.do_train): # use the model itself ref_model = model else: ref_model = create_ref_model(model_args, finetuning_args) # Initialize our Trainer trainer = CustomKTOTrainer( model=model, ref_model=ref_model, args=training_args, finetuning_args=finetuning_args, data_collator=data_collator, callbacks=callbacks, **dataset_module, **tokenizer_module, ) # Training if training_args.do_train: train_result = trainer.train(resume_from_checkpoint=training_args.resume_from_checkpoint) trainer.save_model() trainer.log_metrics("train", train_result.metrics) trainer.save_metrics("train", train_result.metrics) trainer.save_state() if trainer.is_world_process_zero() and finetuning_args.plot_loss: keys = ["loss", "rewards/chosen"] if isinstance(dataset_module.get("eval_dataset"), dict): keys += [f"eval_{key}_loss" for key in dataset_module["eval_dataset"].keys()] else: keys += ["eval_loss"] plot_loss(training_args.output_dir, keys=keys) # Evaluation if training_args.do_eval: metrics = trainer.evaluate(metric_key_prefix="eval") if id(model) == id(ref_model): # unable to compute rewards without a reference model remove_keys = [key for key in metrics.keys() if "rewards" in key] for key in remove_keys: metrics.pop(key) trainer.log_metrics("eval", metrics) trainer.save_metrics("eval", metrics) # Create model card create_modelcard_and_push(trainer, model_args, data_args, training_args, finetuning_args)
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/train/mca/trainer.py
src/llamafactory/train/mca/trainer.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # TODO override the original trainer
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/train/mca/__init__.py
src/llamafactory/train/mca/__init__.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from .workflow import run_dpo, run_pt, run_sft __all__ = ["run_dpo", "run_pt", "run_sft"]
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/train/mca/workflow.py
src/llamafactory/train/mca/workflow.py
# Copyright 2025 the ROLL team and the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import functools from collections.abc import Sequence from copy import deepcopy from typing import TYPE_CHECKING, Any, Optional from transformers import DataCollatorForSeq2Seq from ...data import ( SFTDataCollatorWith4DAttentionMask, get_dataset, get_template_and_fix_tokenizer, ) from ...data.collator import ( PairwiseDataCollatorWithPadding, ) from ...extras.constants import IGNORE_INDEX, MCA_SUPPORTED_MODELS from ...extras.logging import get_logger from ...extras.misc import calculate_tps from ...extras.packages import is_mcore_adapter_available from ...extras.ploting import plot_loss from ...model import load_tokenizer from ..callbacks import SaveProcessorCallback if not is_mcore_adapter_available(): raise ImportError("mcore_adapter is not installed. Please install it with `pip install mcore-adapter`.") from mcore_adapter.models import AutoConfig, AutoModel from mcore_adapter.trainer import DPOTrainer as McaDPOTrainer from mcore_adapter.trainer import McaTrainer from mcore_adapter.trainer.dpo_config import DPOConfig if TYPE_CHECKING: from mcore_adapter.training_args import Seq2SeqTrainingArguments as McaSeq2SeqTrainingArguments from transformers import TrainerCallback from ...hparams import DataArguments, FinetuningArguments, ModelArguments logger = get_logger(__name__) def _data_collator_wrapper(data_collator: Any): @functools.wraps(data_collator) def wrapper(features: Sequence[dict[str, Any]]): labels_key = [k for k in features[0].keys() if k.endswith("labels")] input_ids_key = [k for k in features[0].keys() if k.endswith("input_ids")] for feature in features: if len(labels_key) == 0: # pt feature["labels"] = deepcopy(feature["input_ids"])[1:] for k in labels_key: feature[k] = feature[k][1:] for k in input_ids_key: feature[k] = feature[k][:-1] for k in ["attention_mask", "position_ids"]: if k in feature: feature[k] = feature[k][:-1] return data_collator(features) return wrapper def _check_model_support(model_args: "ModelArguments"): from transformers import AutoConfig as HfAutoConfig config = HfAutoConfig.from_pretrained( model_args.model_name_or_path, trust_remote_code=model_args.trust_remote_code ) if config.model_type not in MCA_SUPPORTED_MODELS: raise ValueError(f"Model {config.model_type} is not supported by MCA.") def run_pt( model_args: "ModelArguments", data_args: "DataArguments", training_args: "McaSeq2SeqTrainingArguments", finetuning_args: "FinetuningArguments", callbacks: Optional[list["TrainerCallback"]] = None, ): tokenizer_module = load_tokenizer(model_args) tokenizer = tokenizer_module["tokenizer"] template = get_template_and_fix_tokenizer(tokenizer, data_args) # dataset needs +1 then cut back due to MCA shift logic data_args.cutoff_len += 1 dataset_module = get_dataset(template, model_args, data_args, training_args, stage="pt", **tokenizer_module) data_args.cutoff_len -= 1 _check_model_support(model_args) model = AutoModel.from_pretrained(model_args.model_name_or_path, training_args) data_collator = DataCollatorForSeq2Seq( tokenizer=tokenizer, pad_to_multiple_of=8, label_pad_token_id=IGNORE_INDEX, ) data_collator = _data_collator_wrapper(data_collator) trainer = McaTrainer( model=model, args=training_args, tokenizer=tokenizer, data_collator=data_collator, callbacks=callbacks, **dataset_module, ) if "processor" in tokenizer_module and tokenizer_module["processor"] is not None: trainer.add_callback(SaveProcessorCallback(tokenizer_module["processor"])) if training_args.do_train: train_result = trainer.train(training_args.resume_from_checkpoint) trainer.save_model() trainer.log_metrics("train", train_result.metrics) trainer.save_metrics("train", train_result.metrics) trainer.save_state() if trainer.is_world_process_zero() and finetuning_args.plot_loss: keys = ["loss"] if isinstance(dataset_module.get("eval_dataset"), dict): keys += [f"eval_{key}_loss" for key in dataset_module["eval_dataset"].keys()] else: keys += ["eval_loss"] plot_loss(training_args.output_dir, keys=keys) def run_sft( model_args: "ModelArguments", data_args: "DataArguments", training_args: "McaSeq2SeqTrainingArguments", finetuning_args: "FinetuningArguments", callbacks: Optional[list["TrainerCallback"]] = None, ): # align packing flags # TODO: FIX SequencePacking data_args.neat_packing = training_args.sequence_packing = data_args.neat_packing or training_args.sequence_packing data_args.packing = data_args.neat_packing or data_args.packing tokenizer_module = load_tokenizer(model_args) tokenizer = tokenizer_module["tokenizer"] template = get_template_and_fix_tokenizer(tokenizer, data_args) # dataset needs +1 then cut back due to MCA shift logic data_args.cutoff_len += 1 dataset_module = get_dataset(template, model_args, data_args, training_args, stage="sft", **tokenizer_module) data_args.cutoff_len -= 1 _check_model_support(model_args) model = AutoModel.from_pretrained(model_args.model_name_or_path, training_args) # optional freezing for qwen2_vl, qwen2_5_vl if getattr(model.config, "hf_model_type", None) in ["qwen2_vl", "qwen2_5_vl", "qwen3_vl"]: params_to_freeze = [] if finetuning_args.freeze_vision_tower: params_to_freeze.extend(["vision_model.blocks", "vision_model.patch_embed"]) if finetuning_args.freeze_multi_modal_projector: params_to_freeze.extend(["multi_modal_projector"]) if finetuning_args.freeze_language_model: params_to_freeze.extend(["embedding", "decoder", "output_layer"]) if params_to_freeze: for name, p in model.named_parameters(): if any(name.startswith(k) for k in params_to_freeze): p.requires_grad_(False) pad_to_max = training_args.expert_model_parallel_size is not None and training_args.expert_model_parallel_size > 1 data_collator = SFTDataCollatorWith4DAttentionMask( template=template, padding="max_length" if pad_to_max else "longest", max_length=data_args.cutoff_len if pad_to_max else None, pad_to_multiple_of=64, label_pad_token_id=IGNORE_INDEX, **tokenizer_module, ) data_collator = _data_collator_wrapper(data_collator) trainer = McaTrainer( model=model, args=training_args, tokenizer=tokenizer, data_collator=data_collator, callbacks=callbacks, **dataset_module, ) if "processor" in tokenizer_module and tokenizer_module["processor"] is not None: trainer.add_callback(SaveProcessorCallback(tokenizer_module["processor"])) train_result = trainer.train(training_args.resume_from_checkpoint) trainer.save_model() trainer.log_metrics("train", train_result.metrics) trainer.save_metrics("train", train_result.metrics) trainer.save_state() if trainer.is_world_process_zero() and finetuning_args.plot_loss: keys = ["loss"] if isinstance(dataset_module.get("eval_dataset"), dict): keys += [f"eval_{key}_loss" for key in dataset_module["eval_dataset"].keys()] else: keys += ["eval_loss"] plot_loss(training_args.output_dir, keys=keys) def run_dpo( model_args: "ModelArguments", data_args: "DataArguments", training_args: "McaSeq2SeqTrainingArguments", finetuning_args: "FinetuningArguments", callbacks: Optional[list["TrainerCallback"]] = None, ): tokenizer_module = load_tokenizer(model_args) tokenizer = tokenizer_module["tokenizer"] template = get_template_and_fix_tokenizer(tokenizer, data_args) _check_model_support(model_args) model = AutoModel.from_pretrained(model_args.model_name_or_path, training_args) if finetuning_args.use_ref_model: ref_config = AutoConfig.from_pretrained(model_args.model_name_or_path, training_args) ref_model = AutoModel.from_config(ref_config) ref_model.load_state_dict(model.state_dict()) else: ref_model = None # dataset needs +1 then cut back due to MCA shift logic data_args.cutoff_len += 1 dataset_module = get_dataset(template, model_args, data_args, training_args, stage="rm", **tokenizer_module) data_args.cutoff_len -= 1 pad_to_max = training_args.expert_model_parallel_size is not None and training_args.expert_model_parallel_size > 1 dpo_config = DPOConfig( beta=finetuning_args.pref_beta, pref_loss=finetuning_args.pref_loss, label_smoothing=finetuning_args.dpo_label_smoothing, ) data_collator = PairwiseDataCollatorWithPadding( template=template, pad_to_multiple_of=64, padding="max_length" if pad_to_max else "longest", max_length=data_args.cutoff_len if pad_to_max else None, label_pad_token_id=IGNORE_INDEX, **tokenizer_module, ) data_collator = _data_collator_wrapper(data_collator) trainer = McaDPOTrainer( model=model, ref_model=ref_model, args=training_args, train_config=dpo_config, tokenizer=tokenizer, data_collator=data_collator, callbacks=callbacks, **dataset_module, ) if "processor" in tokenizer_module and tokenizer_module["processor"] is not None: trainer.add_callback(SaveProcessorCallback(tokenizer_module["processor"])) train_result = trainer.train(training_args.resume_from_checkpoint) trainer.save_model() if finetuning_args.include_effective_tokens_per_second: train_result.metrics["effective_tokens_per_sec"] = calculate_tps( dataset_module["train_dataset"], train_result.metrics, stage="rm" ) trainer.log_metrics("train", train_result.metrics) trainer.save_metrics("train", train_result.metrics) trainer.save_state() if trainer.is_world_process_zero() and finetuning_args.plot_loss: keys = ["loss", "rewards/accuracies"] if isinstance(dataset_module.get("eval_dataset"), dict): keys += [f"eval_{key}_loss" for key in dataset_module["eval_dataset"].keys()] else: keys += ["eval_loss"] plot_loss(training_args.output_dir, keys=keys)
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/train/rm/trainer.py
src/llamafactory/train/rm/trainer.py
# Copyright 2025 HuggingFace Inc. and the LlamaFactory team. # # This code is inspired by the HuggingFace's transformers library. # https://github.com/huggingface/transformers/blob/v4.40.0/src/transformers/trainer.py # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import json import os from types import MethodType from typing import TYPE_CHECKING, Optional, Union import torch from transformers import Trainer from typing_extensions import override from ...extras import logging from ...extras.packages import is_transformers_version_greater_than from ..callbacks import FixValueHeadModelCallback, SaveProcessorCallback from ..trainer_utils import create_custom_optimizer, create_custom_scheduler if TYPE_CHECKING: from transformers import PreTrainedModel, ProcessorMixin from transformers.trainer import PredictionOutput from ...hparams import FinetuningArguments logger = logging.get_logger(__name__) class PairwiseTrainer(Trainer): r"""Inherits Trainer to compute pairwise loss.""" def __init__( self, finetuning_args: "FinetuningArguments", processor: Optional["ProcessorMixin"], **kwargs ) -> None: if is_transformers_version_greater_than("4.46"): kwargs["processing_class"] = kwargs.pop("tokenizer") super().__init__(**kwargs) self.model_accepts_loss_kwargs = False # overwrite trainer's default behavior self.finetuning_args = finetuning_args self.can_return_loss = True # override property to return eval_loss self.add_callback(FixValueHeadModelCallback) if processor is not None: self.add_callback(SaveProcessorCallback(processor)) if finetuning_args.use_badam: from badam import BAdamCallback, clip_grad_norm_old_version # type: ignore self.accelerator.clip_grad_norm_ = MethodType(clip_grad_norm_old_version, self.accelerator) self.add_callback(BAdamCallback) @override def create_optimizer(self) -> "torch.optim.Optimizer": if self.optimizer is None: self.optimizer = create_custom_optimizer(self.model, self.args, self.finetuning_args) return super().create_optimizer() @override def create_scheduler( self, num_training_steps: int, optimizer: Optional["torch.optim.Optimizer"] = None ) -> "torch.optim.lr_scheduler.LRScheduler": create_custom_scheduler(self.args, num_training_steps, optimizer) return super().create_scheduler(num_training_steps, optimizer) @override def _get_train_sampler(self, *args, **kwargs) -> Optional["torch.utils.data.Sampler"]: if self.finetuning_args.disable_shuffling: return torch.utils.data.SequentialSampler(self.train_dataset) return super()._get_train_sampler(*args, **kwargs) @override def compute_loss( self, model: "PreTrainedModel", inputs: dict[str, "torch.Tensor"], return_outputs: bool = False, **kwargs ) -> Union["torch.Tensor", tuple["torch.Tensor", list["torch.Tensor"]]]: r"""Compute pairwise loss. The first n examples are chosen and the last n examples are rejected. Subclass and override to inject custom behavior. Note that the first element will be removed from the output tuple. See: https://github.com/huggingface/transformers/blob/v4.40.0/src/transformers/trainer.py#L3842 """ _, _, values = model(**inputs, output_hidden_states=True, return_dict=True, use_cache=False) batch_size = inputs["input_ids"].size(0) // 2 chosen_masks, rejected_masks = torch.split(inputs["attention_mask"], batch_size, dim=0) chosen_rewards, rejected_rewards = torch.split(values, batch_size, dim=0) chosen_scores = chosen_rewards.gather(dim=-1, index=(chosen_masks.sum(dim=-1, keepdim=True) - 1)) rejected_scores = rejected_rewards.gather(dim=-1, index=(rejected_masks.sum(dim=-1, keepdim=True) - 1)) chosen_scores, rejected_scores = chosen_scores.squeeze(), rejected_scores.squeeze() loss = -torch.nn.functional.logsigmoid(chosen_scores.float() - rejected_scores.float()).mean() if return_outputs: return loss, (loss, chosen_scores, rejected_scores) else: return loss def save_predictions(self, predict_results: "PredictionOutput") -> None: r"""Save model predictions to `output_dir`. A custom behavior that not contained in Seq2SeqTrainer. """ if not self.is_world_process_zero(): return output_prediction_file = os.path.join(self.args.output_dir, "generated_predictions.jsonl") logger.info_rank0(f"Saving prediction results to {output_prediction_file}") chosen_scores, rejected_scores = predict_results.predictions with open(output_prediction_file, "w", encoding="utf-8") as writer: res: list[str] = [] for c_score, r_score in zip(chosen_scores, rejected_scores): res.append(json.dumps({"chosen": round(float(c_score), 2), "rejected": round(float(r_score), 2)})) writer.write("\n".join(res))
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/train/rm/__init__.py
src/llamafactory/train/rm/__init__.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from .workflow import run_rm __all__ = ["run_rm"]
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/train/rm/workflow.py
src/llamafactory/train/rm/workflow.py
# Copyright 2025 HuggingFace Inc. and the LlamaFactory team. # # This code is inspired by the HuggingFace's transformers library. # https://github.com/huggingface/transformers/blob/v4.40.0/examples/pytorch/summarization/run_summarization.py # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING, Optional from ...data import PairwiseDataCollatorWithPadding, get_dataset, get_template_and_fix_tokenizer from ...extras.ploting import plot_loss from ...model import load_model, load_tokenizer from ..callbacks import fix_valuehead_checkpoint from ..trainer_utils import create_modelcard_and_push from .metric import ComputeAccuracy from .trainer import PairwiseTrainer if TYPE_CHECKING: from transformers import Seq2SeqTrainingArguments, TrainerCallback from ...hparams import DataArguments, FinetuningArguments, ModelArguments def run_rm( model_args: "ModelArguments", data_args: "DataArguments", training_args: "Seq2SeqTrainingArguments", finetuning_args: "FinetuningArguments", callbacks: Optional[list["TrainerCallback"]] = None, ): tokenizer_module = load_tokenizer(model_args) tokenizer = tokenizer_module["tokenizer"] template = get_template_and_fix_tokenizer(tokenizer, data_args) dataset_module = get_dataset(template, model_args, data_args, training_args, stage="rm", **tokenizer_module) model = load_model(tokenizer, model_args, finetuning_args, training_args.do_train, add_valuehead=True) data_collator = PairwiseDataCollatorWithPadding( template=template, model=model, pad_to_multiple_of=8, **tokenizer_module ) # Initialize our Trainer trainer = PairwiseTrainer( model=model, args=training_args, finetuning_args=finetuning_args, data_collator=data_collator, callbacks=callbacks, compute_metrics=ComputeAccuracy(), **dataset_module, **tokenizer_module, ) # Training if training_args.do_train: train_result = trainer.train(resume_from_checkpoint=training_args.resume_from_checkpoint) trainer.save_model() if training_args.should_save: fix_valuehead_checkpoint(model, training_args.output_dir, training_args.save_safetensors) trainer.log_metrics("train", train_result.metrics) trainer.save_metrics("train", train_result.metrics) trainer.save_state() if trainer.is_world_process_zero() and finetuning_args.plot_loss: keys = ["loss"] if isinstance(dataset_module.get("eval_dataset"), dict): keys += sum( [[f"eval_{key}_loss", f"eval_{key}_accuracy"] for key in dataset_module["eval_dataset"].keys()], [] ) else: keys += ["eval_loss", "eval_accuracy"] plot_loss(training_args.output_dir, keys=keys) # Evaluation if training_args.do_eval: metrics = trainer.evaluate(metric_key_prefix="eval") trainer.log_metrics("eval", metrics) trainer.save_metrics("eval", metrics) # Predict if training_args.do_predict: predict_results = trainer.predict(dataset_module["eval_dataset"], metric_key_prefix="predict") trainer.log_metrics("predict", predict_results.metrics) trainer.save_metrics("predict", predict_results.metrics) trainer.save_predictions(predict_results) # Create model card create_modelcard_and_push(trainer, model_args, data_args, training_args, finetuning_args)
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/train/rm/metric.py
src/llamafactory/train/rm/metric.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from dataclasses import dataclass from typing import TYPE_CHECKING import numpy as np from ...extras.misc import numpify if TYPE_CHECKING: from transformers import EvalPrediction @dataclass class ComputeAccuracy: r"""Compute reward accuracy and support `batch_eval_metrics`.""" def _dump(self) -> dict[str, float] | None: result = None if hasattr(self, "score_dict"): result = {k: float(np.mean(v)) for k, v in self.score_dict.items()} self.score_dict = {"accuracy": []} return result def __post_init__(self): self._dump() def __call__(self, eval_preds: "EvalPrediction", compute_result: bool = True) -> dict[str, float] | None: chosen_scores, rejected_scores = numpify(eval_preds.predictions[0]), numpify(eval_preds.predictions[1]) if not chosen_scores.shape: self.score_dict["accuracy"].append(chosen_scores > rejected_scores) else: for i in range(len(chosen_scores)): self.score_dict["accuracy"].append(chosen_scores[i] > rejected_scores[i]) if compute_result: return self._dump()
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/model/patcher.py
src/llamafactory/model/patcher.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from types import MethodType from typing import TYPE_CHECKING, Any import torch from peft import PeftModel from transformers import GenerationMixin, PreTrainedModel, PreTrainedTokenizerBase from transformers.integrations import is_deepspeed_zero3_enabled from transformers.modeling_utils import is_fsdp_enabled from ..extras import logging from ..extras.misc import infer_optim_dtype from ..extras.packages import is_transformers_version_greater_than from .model_utils.attention import configure_attn_implementation, print_attn_implementation from .model_utils.checkpointing import prepare_model_for_training from .model_utils.embedding import resize_embedding_layer from .model_utils.kv_cache import configure_kv_cache from .model_utils.longlora import configure_longlora from .model_utils.moe import add_z3_leaf_module, configure_moe from .model_utils.packing import configure_packing from .model_utils.quantization import configure_quantization from .model_utils.rope import configure_rope from .model_utils.valuehead import prepare_valuehead_model from .model_utils.visual import autocast_projector_dtype, configure_visual_model if TYPE_CHECKING: from transformers import PretrainedConfig, PreTrainedTokenizer, ProcessorMixin from trl import AutoModelForCausalLMWithValueHead from ..hparams import ModelArguments if is_transformers_version_greater_than("4.57.0"): from transformers.models.qwen3_omni_moe import modeling_qwen3_omni_moe logger = logging.get_logger(__name__) def patch_qwen3_omni_moe_thinker_text_sparse_moe_block(): if is_transformers_version_greater_than("4.57.0") and not is_transformers_version_greater_than("4.58.0"): from .model_utils.moe import Qwen3OmniMoeThinkerTextSparseMoeBlock logger.warning_rank0( "You are using transformers with 4.x version, the Qwen3OmniMoeThinkerTextSparseMoeBlock will have some issues about deepspeed zero2 and fsdp2 training, so that we patched this model to avoid it. Transformers v5.0.0rc0 has fixed the issue, you can also try to update the transformers to using qwen3_omni. See more information on https://github.com/hiyouga/LLaMA-Factory/issues/9628." ) modeling_qwen3_omni_moe.Qwen3OmniMoeThinkerTextSparseMoeBlock = Qwen3OmniMoeThinkerTextSparseMoeBlock def patch_tokenizer(tokenizer: "PreTrainedTokenizer", model_args: "ModelArguments") -> None: if "PreTrainedTokenizerBase" not in str(tokenizer._pad.__func__): tokenizer._pad = MethodType(PreTrainedTokenizerBase._pad, tokenizer) if model_args.model_max_length is not None and tokenizer.model_max_length < model_args.model_max_length: tokenizer.model_max_length = model_args.model_max_length # enlarge the tokenizer max length if model_args.add_tokens is not None: num_added_tokens = tokenizer.add_tokens(new_tokens=model_args.add_tokens, special_tokens=False) logger.info_rank0("Add tokens {} to tokenizer's vocabulary.".format(",".join(model_args.add_tokens))) if num_added_tokens > 0 and not model_args.resize_vocab: model_args.resize_vocab = True logger.warning_rank0("New tokens have been added, changed `resize_vocab` to True.") if model_args.add_special_tokens is not None: num_added_special_tokens = tokenizer.add_tokens(new_tokens=model_args.add_special_tokens, special_tokens=True) logger.info_rank0( "Add special tokens {} to tokenizer's vocabulary.".format(",".join(model_args.add_special_tokens)) ) if num_added_special_tokens > 0 and not model_args.resize_vocab: model_args.resize_vocab = True logger.warning_rank0("New special tokens have been added, changed `resize_vocab` to True.") def patch_processor( processor: "ProcessorMixin", tokenizer: "PreTrainedTokenizer", model_args: "ModelArguments", ) -> None: setattr(processor, "tokenizer", tokenizer) setattr(processor, "image_max_pixels", model_args.image_max_pixels) setattr(processor, "image_min_pixels", model_args.image_min_pixels) setattr(processor, "image_do_pan_and_scan", model_args.image_do_pan_and_scan) setattr(processor, "crop_to_patches", model_args.crop_to_patches) setattr(processor, "video_max_pixels", model_args.video_max_pixels) setattr(processor, "video_min_pixels", model_args.video_min_pixels) setattr(processor, "video_fps", model_args.video_fps) setattr(processor, "video_maxlen", model_args.video_maxlen) setattr(processor, "use_audio_in_video", model_args.use_audio_in_video) setattr(processor, "audio_sampling_rate", model_args.audio_sampling_rate) def patch_config( config: "PretrainedConfig", tokenizer: "PreTrainedTokenizer", model_args: "ModelArguments", init_kwargs: dict[str, Any], is_trainable: bool, ) -> None: if model_args.compute_dtype is None: # priority: bf16 > fp16 > fp32 if model_args.infer_dtype != "auto" and not is_trainable: model_args.compute_dtype = getattr(torch, model_args.infer_dtype) else: model_args.compute_dtype = infer_optim_dtype(model_dtype=getattr(config, "torch_dtype", None)) configure_attn_implementation(config, model_args) configure_rope(config, model_args) configure_longlora(config, model_args, is_trainable) configure_quantization(config, tokenizer, model_args, is_trainable, init_kwargs) configure_moe(config, model_args, is_trainable) configure_visual_model(config) configure_packing(model_args, is_trainable) configure_kv_cache(config, model_args, is_trainable) if getattr(config, "model_type", None) == "qwen": setattr(config, "use_flash_attn", model_args.flash_attn == "fa2") for dtype_name, dtype in [("fp16", torch.float16), ("bf16", torch.bfloat16), ("fp32", torch.float32)]: setattr(config, dtype_name, model_args.compute_dtype == dtype) if getattr(config, "model_type", None) == "minicpmo": setattr(config, "init_audio", True) setattr(config, "init_tts", False) # replace the top-k gating method if getattr(config, "model_type", None) == "kimi_vl" and is_trainable: setattr(config.text_config, "topk_method", "greedy") architectures = getattr(config, "architectures", None) if isinstance(architectures, list) and "InternVLChatModel" in architectures: raise ValueError( "Please download the internvl models in a Hugging Face–compatible format " "(for example, https://huggingface.co/OpenGVLab/InternVL3-8B-hf)." ) if isinstance(architectures, list) and "LlavaLlamaForCausalLM" in architectures: raise ValueError("Please download llava models with hf-compatible format: https://huggingface.co/llava-hf") if getattr(config, "model_type", None) == "internlm3" and not is_transformers_version_greater_than("4.47.1"): raise RuntimeError("InternLM3 model requires transformers>=4.47.1, please upgrade it.") if getattr(config, "model_type", None) == "qwen3_omni_moe": patch_qwen3_omni_moe_thinker_text_sparse_moe_block() # deepspeed zero3 is not compatible with low_cpu_mem_usage init_kwargs["low_cpu_mem_usage"] = model_args.low_cpu_mem_usage and (not is_deepspeed_zero3_enabled()) # fsdp/deepspeed zero3 does not need device map if not (is_deepspeed_zero3_enabled() or is_fsdp_enabled()) and init_kwargs["low_cpu_mem_usage"]: if "device_map" not in init_kwargs and model_args.device_map: init_kwargs["device_map"] = model_args.device_map # device map requires low_cpu_mem_usage=True if init_kwargs.get("device_map", None) == "auto": init_kwargs["offload_folder"] = model_args.offload_folder def patch_model( model: "PreTrainedModel", tokenizer: "PreTrainedTokenizer", model_args: "ModelArguments", is_trainable: bool, add_valuehead: bool, ) -> None: gen_config = model.generation_config # check and fix generation config if not gen_config.do_sample and ( (gen_config.temperature is not None and gen_config.temperature != 1.0) or (gen_config.top_p is not None and gen_config.top_p != 1.0) or (gen_config.typical_p is not None and gen_config.typical_p != 1.0) ): gen_config.do_sample = True if getattr(model.config, "model_type", None) not in ["minicpmv", "minicpmo"] and "GenerationMixin" not in str( model.generate.__func__ ): model.generate = MethodType(GenerationMixin.generate, model) if add_valuehead: prepare_valuehead_model(model) if model_args.resize_vocab: resize_embedding_layer( model, tokenizer, new_special_tokens_config=getattr(model_args, "_special_token_descriptions", None), init_special_tokens=model_args.init_special_tokens, ) if is_trainable: if getattr(model.config, "model_type", None) == "gemma3n": setattr(model_args, "disable_gradient_checkpointing", True) prepare_model_for_training(model, model_args) autocast_projector_dtype(model, model_args) add_z3_leaf_module(model) if not model_args.use_unsloth: print_attn_implementation(model.config) try: model.add_model_tags(["llama-factory"]) except Exception: logger.warning_rank0("Cannot properly tag the model.") def patch_valuehead_model(model: "AutoModelForCausalLMWithValueHead") -> None: def tie_weights(self: "AutoModelForCausalLMWithValueHead") -> None: if isinstance(self.pretrained_model, PreTrainedModel): self.pretrained_model.tie_weights() def get_input_embeddings(self: "AutoModelForCausalLMWithValueHead") -> torch.nn.Module: if isinstance(self.pretrained_model, PreTrainedModel): return self.pretrained_model.get_input_embeddings() def get_output_embeddings(self: "AutoModelForCausalLMWithValueHead") -> torch.nn.Module: if isinstance(self.pretrained_model, PreTrainedModel): return self.pretrained_model.get_output_embeddings() def create_or_update_model_card(self: "AutoModelForCausalLMWithValueHead", output_dir: str) -> None: if isinstance(self.pretrained_model, PeftModel): self.pretrained_model.create_or_update_model_card(output_dir) def get_rope_index_func(self: "AutoModelForCausalLMWithValueHead"): if isinstance(self.pretrained_model, PeftModel): base_model = self.pretrained_model.base_model.model else: base_model = self.pretrained_model if base_model and hasattr(base_model, "get_rope_index"): return base_model.get_rope_index elif base_model and hasattr(base_model, "model") and hasattr(base_model.model, "get_rope_index"): return base_model.model.get_rope_index else: return None ignore_modules = [name for name, _ in model.named_parameters() if "pretrained_model" in name] setattr(model, "_keys_to_ignore_on_save", ignore_modules) setattr(model, "tie_weights", MethodType(tie_weights, model)) setattr(model, "get_input_embeddings", MethodType(get_input_embeddings, model)) setattr(model, "get_output_embeddings", MethodType(get_output_embeddings, model)) setattr(model, "get_rope_index", get_rope_index_func(model)) setattr(model, "create_or_update_model_card", MethodType(create_or_update_model_card, model))
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/model/loader.py
src/llamafactory/model/loader.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import os from typing import TYPE_CHECKING, Any, Optional, TypedDict from transformers import ( AutoConfig, AutoModelForCausalLM, AutoModelForImageTextToText, AutoModelForSeq2SeqLM, AutoModelForTextToWaveform, AutoModelForVision2Seq, AutoProcessor, AutoTokenizer, ) from trl import AutoModelForCausalLMWithValueHead from ..extras import logging from ..extras.misc import count_parameters, skip_check_imports, try_download_model_from_other_hub from .adapter import init_adapter from .model_utils.ktransformers import load_kt_pretrained_model from .model_utils.liger_kernel import apply_liger_kernel from .model_utils.misc import register_autoclass from .model_utils.mod import convert_pretrained_model_to_mod, load_mod_pretrained_model from .model_utils.unsloth import load_unsloth_pretrained_model from .model_utils.valuehead import load_valuehead_params from .patcher import patch_config, patch_model, patch_processor, patch_tokenizer, patch_valuehead_model if TYPE_CHECKING: from transformers import PretrainedConfig, PreTrainedModel, PreTrainedTokenizer, ProcessorMixin from ..hparams import FinetuningArguments, ModelArguments logger = logging.get_logger(__name__) class TokenizerModule(TypedDict): tokenizer: "PreTrainedTokenizer" processor: Optional["ProcessorMixin"] def _get_init_kwargs(model_args: "ModelArguments") -> dict[str, Any]: r"""Get arguments to load config/tokenizer/model. Note: including inplace operation of model_args. """ skip_check_imports() model_args.model_name_or_path = try_download_model_from_other_hub(model_args) return { "trust_remote_code": model_args.trust_remote_code, "cache_dir": model_args.cache_dir, "revision": model_args.model_revision, "token": model_args.hf_hub_token, } def load_tokenizer(model_args: "ModelArguments") -> "TokenizerModule": r"""Load pretrained tokenizer and optionally loads processor. Note: including inplace operation of model_args. """ init_kwargs = _get_init_kwargs(model_args) try: tokenizer = AutoTokenizer.from_pretrained( model_args.model_name_or_path, use_fast=model_args.use_fast_tokenizer, split_special_tokens=model_args.split_special_tokens, padding_side="right", **init_kwargs, ) except ValueError: # try another one tokenizer = AutoTokenizer.from_pretrained( model_args.model_name_or_path, use_fast=not model_args.use_fast_tokenizer, padding_side="right", **init_kwargs, ) except Exception as e: raise OSError("Failed to load tokenizer.") from e patch_tokenizer(tokenizer, model_args) try: processor = AutoProcessor.from_pretrained( model_args.model_name_or_path, use_fast=model_args.use_fast_tokenizer, **init_kwargs, ) except ValueError: # try another one processor = AutoProcessor.from_pretrained( model_args.model_name_or_path, use_fast=not model_args.use_fast_tokenizer, **init_kwargs, ) except Exception as e: logger.info_rank0(f"Failed to load processor: {e}.") processor = None # Avoid load tokenizer, see: # https://github.com/huggingface/transformers/blob/v4.40.0/src/transformers/models/auto/processing_auto.py#L324 if processor is not None and "Processor" not in processor.__class__.__name__: logger.debug("The loaded processor is not an instance of Processor. Dropping it.") processor = None if processor is not None: patch_processor(processor, tokenizer, model_args) return {"tokenizer": tokenizer, "processor": processor} def load_config(model_args: "ModelArguments") -> "PretrainedConfig": r"""Load model config.""" init_kwargs = _get_init_kwargs(model_args) return AutoConfig.from_pretrained(model_args.model_name_or_path, **init_kwargs) def load_model( tokenizer: "PreTrainedTokenizer", model_args: "ModelArguments", finetuning_args: "FinetuningArguments", is_trainable: bool = False, add_valuehead: bool = False, ) -> "PreTrainedModel": r"""Load pretrained model.""" init_kwargs = _get_init_kwargs(model_args) config = load_config(model_args) patch_config(config, tokenizer, model_args, init_kwargs, is_trainable) apply_liger_kernel(config, model_args, is_trainable, require_logits=(finetuning_args.stage not in ["pt", "sft"])) model = None lazy_load = False if model_args.use_kt: from ktransformers.sft.monkey_patch_torch_module import install_patch install_patch() model = load_kt_pretrained_model(config, model_args) elif model_args.use_unsloth: if model_args.adapter_name_or_path is not None: lazy_load = True elif is_trainable: model = load_unsloth_pretrained_model(config, model_args, finetuning_args) if model is None and not lazy_load: init_kwargs["config"] = config init_kwargs["pretrained_model_name_or_path"] = model_args.model_name_or_path init_kwargs["torch_dtype"] = "auto" if model_args.mixture_of_depths == "load": model = load_mod_pretrained_model(**init_kwargs) else: if type(config) in AutoModelForImageTextToText._model_mapping.keys(): # image-text load_class = AutoModelForImageTextToText elif type(config) in AutoModelForVision2Seq._model_mapping.keys(): # image-text load_class = AutoModelForVision2Seq elif type(config) in AutoModelForSeq2SeqLM._model_mapping.keys(): # audio-text load_class = AutoModelForSeq2SeqLM elif type(config) in AutoModelForTextToWaveform._model_mapping.keys(): # audio hack for qwen omni load_class = AutoModelForTextToWaveform else: load_class = AutoModelForCausalLM if model_args.train_from_scratch: model = load_class.from_config(config, trust_remote_code=model_args.trust_remote_code) else: model = load_class.from_pretrained(**init_kwargs) if getattr(model.config, "model_type", None) in ["qwen2_5_omni", "qwen3_omni_moe"]: model = getattr(model, "thinker") if model_args.mixture_of_depths == "convert": model = convert_pretrained_model_to_mod(model, config, model_args) if not lazy_load: patch_model(model, tokenizer, model_args, is_trainable, add_valuehead) register_autoclass(config, model, tokenizer) model = init_adapter(config, model, model_args, finetuning_args, is_trainable) if add_valuehead: model = AutoModelForCausalLMWithValueHead.from_pretrained(model) patch_valuehead_model(model) if model_args.adapter_name_or_path is not None: vhead_path = model_args.adapter_name_or_path[-1] else: vhead_path = model_args.model_name_or_path vhead_params = load_valuehead_params(vhead_path, model_args) if vhead_params is not None: model.load_state_dict(vhead_params, strict=False) logger.info_rank0(f"Loaded valuehead from checkpoint: {vhead_path}") if not is_trainable: model.requires_grad_(False) model.eval() else: model.train() # Borrowing the kernel plugins ability of v1 to temporarily apply the NPU fusion operator to v0, # it is turned off by default, and can be discarded after the transition period ends. if model_args.use_v1_kernels and is_trainable: logger.warning_rank0( "You are try to using future feature about kernels, please note that this feature " "is not supported for all models. If get any error, please disable this feature, or report the issue." ) from ..v1.plugins.model_plugins.kernels.interface import apply_default_kernels model = apply_default_kernels(model=model, include_kernels=model_args.use_v1_kernels) trainable_params, all_param = count_parameters(model) if is_trainable: param_stats = ( f"trainable params: {trainable_params:,} || " f"all params: {all_param:,} || trainable%: {100 * trainable_params / all_param:.4f}" ) else: param_stats = f"all params: {all_param:,}" logger.info_rank0(param_stats) if model_args.print_param_status and int(os.getenv("LOCAL_RANK", "0")) == 0: for name, param in model.named_parameters(): print(f"name: {name}, dtype: {param.dtype}, device: {param.device}, trainable: {param.requires_grad}") return model
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/model/__init__.py
src/llamafactory/model/__init__.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from .loader import load_config, load_model, load_tokenizer from .model_utils.misc import find_all_linear_modules from .model_utils.quantization import QuantizationMethod from .model_utils.valuehead import load_valuehead_params __all__ = [ "QuantizationMethod", "find_all_linear_modules", "load_config", "load_model", "load_tokenizer", "load_valuehead_params", ]
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/model/adapter.py
src/llamafactory/model/adapter.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from typing import TYPE_CHECKING import torch from peft import LoraConfig, LoraModel, OFTConfig, PeftModel, TaskType, get_peft_model from transformers.integrations import is_deepspeed_zero3_enabled from ..extras import logging from ..extras.constants import EngineName from .model_utils.ktransformers import get_kt_peft_model, load_kt_peft_model from .model_utils.misc import find_all_linear_modules, find_expanded_modules from .model_utils.quantization import QuantizationMethod from .model_utils.unsloth import get_unsloth_peft_model, load_unsloth_peft_model from .model_utils.visual import COMPOSITE_MODELS, get_forbidden_modules, patch_target_modules if TYPE_CHECKING: from transformers import PretrainedConfig, PreTrainedModel from ..hparams import FinetuningArguments, ModelArguments logger = logging.get_logger(__name__) def _setup_full_tuning( model: "PreTrainedModel", finetuning_args: "FinetuningArguments", is_trainable: bool, cast_trainable_params_to_fp32: bool, ) -> None: if not is_trainable: return logger.info_rank0("Fine-tuning method: Full") forbidden_modules = get_forbidden_modules(model.config, finetuning_args) for name, param in model.named_parameters(): if not any(forbidden_module in name for forbidden_module in forbidden_modules): if cast_trainable_params_to_fp32: param.data = param.data.to(torch.float32) else: param.requires_grad_(False) def _setup_freeze_tuning( model: "PreTrainedModel", finetuning_args: "FinetuningArguments", is_trainable: bool, cast_trainable_params_to_fp32: bool, ) -> None: if not is_trainable: return logger.info_rank0("Fine-tuning method: Freeze") if hasattr(model.config, "text_config"): # composite models config = getattr(model.config, "text_config") else: config = model.config num_layers = ( getattr(config, "num_hidden_layers", None) or getattr(config, "num_layers", None) or getattr(config, "n_layer", None) ) if not num_layers: raise ValueError("Current model does not support freeze tuning.") if finetuning_args.use_llama_pro: if num_layers % finetuning_args.freeze_trainable_layers != 0: raise ValueError( f"`num_layers` {num_layers} should be " f"divisible by `num_layer_trainable` {finetuning_args.freeze_trainable_layers}." ) stride = num_layers // finetuning_args.freeze_trainable_layers trainable_layer_ids = range(stride - 1, num_layers + stride - 1, stride) elif finetuning_args.freeze_trainable_layers > 0: # fine-tuning the last n layers if num_layer_trainable > 0 trainable_layer_ids = range(max(0, num_layers - finetuning_args.freeze_trainable_layers), num_layers) else: # fine-tuning the first n layers if num_layer_trainable < 0 trainable_layer_ids = range(min(-finetuning_args.freeze_trainable_layers, num_layers)) hidden_modules = set() non_hidden_modules = set() for name, _ in model.named_parameters(): if ".0." in name: hidden_modules.add(name.split(".0.")[-1].split(".")[0]) elif ".1." in name: # MoD starts from layer 1 hidden_modules.add(name.split(".1.")[-1].split(".")[0]) if re.search(r"\.\d+\.", name) is None: non_hidden_modules.add(name.split(".")[-2]) # remove weight/bias trainable_layers = [] for module_name in finetuning_args.freeze_trainable_modules: if module_name != "all" and module_name not in hidden_modules: raise ValueError( "Module {} is not found, please choose from {}".format(module_name, ", ".join(hidden_modules)) ) for idx in trainable_layer_ids: trainable_layers.append(".{:d}.{}".format(idx, module_name if module_name != "all" else "")) if finetuning_args.freeze_extra_modules: for module_name in finetuning_args.freeze_extra_modules: if module_name not in non_hidden_modules: raise ValueError( "Module {} is not found, please choose from {}".format(module_name, ", ".join(non_hidden_modules)) ) trainable_layers.append(module_name) model_type = getattr(model.config, "model_type", None) if not finetuning_args.freeze_multi_modal_projector and model_type in COMPOSITE_MODELS: trainable_layers.append(COMPOSITE_MODELS[model_type].projector_key) forbidden_modules = get_forbidden_modules(model.config, finetuning_args) for name, param in model.named_parameters(): if any(trainable_layer in name for trainable_layer in trainable_layers) and not any( forbidden_module in name for forbidden_module in forbidden_modules ): if cast_trainable_params_to_fp32: param.data = param.data.to(torch.float32) else: param.requires_grad_(False) logger.info_rank0("Set trainable layers: {}".format(",".join(trainable_layers))) def _setup_lora_tuning( config: "PretrainedConfig", model: "PreTrainedModel", model_args: "ModelArguments", finetuning_args: "FinetuningArguments", is_trainable: bool, cast_trainable_params_to_fp32: bool, ) -> "PeftModel": if is_trainable: if finetuning_args.finetuning_type == "oft": logger.info_rank0("Fine-tuning method: OFT") else: logger.info_rank0("Fine-tuning method: {}".format("DoRA" if finetuning_args.use_dora else "LoRA")) adapter_to_resume = None if model_args.adapter_name_or_path is not None: is_mergeable = True if getattr(model, "quantization_method", None): # merge lora in quantized model is unstable assert len(model_args.adapter_name_or_path) == 1, "Quantized model only accepts a single adapter." is_mergeable = False if is_deepspeed_zero3_enabled(): assert len(model_args.adapter_name_or_path) == 1, "Cannot use multiple adapters in DeepSpeed ZeRO-3." is_mergeable = False if model_args.use_kt: assert len(model_args.adapter_name_or_path) == 1, "KTransformers model only accepts a single adapter" is_mergeable = False if model_args.use_unsloth: assert len(model_args.adapter_name_or_path) == 1, "Unsloth model only accepts a single adapter." is_mergeable = False if (is_trainable and not finetuning_args.create_new_adapter) or (not is_mergeable): adapter_to_merge = model_args.adapter_name_or_path[:-1] adapter_to_resume = model_args.adapter_name_or_path[-1] else: adapter_to_merge = model_args.adapter_name_or_path init_kwargs = { "subfolder": model_args.adapter_folder, "offload_folder": model_args.offload_folder, "cache_dir": model_args.cache_dir, "revision": model_args.model_revision, "token": model_args.hf_hub_token, } if model_args.use_kt: if model_args.infer_backend != EngineName.KT: raise ValueError( "We should use ktransformers as backend to infer the adapter fine-tuned by ktransformers." ) for adapter in adapter_to_merge: model: LoraModel = PeftModel.from_pretrained(model, adapter, **init_kwargs) model = model.merge_and_unload() if len(adapter_to_merge) > 0: logger.info_rank0(f"Merged {len(adapter_to_merge)} adapter(s).") if adapter_to_resume is not None: # resume lora training if model_args.use_kt: model = load_kt_peft_model(model_args, model) elif model_args.use_unsloth: model = load_unsloth_peft_model(config, model_args, finetuning_args, is_trainable=is_trainable) else: model = PeftModel.from_pretrained(model, adapter_to_resume, is_trainable=is_trainable, **init_kwargs) logger.info_rank0("Loaded adapter(s): {}".format(",".join(model_args.adapter_name_or_path))) if is_trainable and adapter_to_resume is None: # create new lora weights while training if len(finetuning_args.lora_target) == 1 and finetuning_args.lora_target[0] == "all": target_modules = find_all_linear_modules(model, finetuning_args.freeze_vision_tower) else: target_modules = finetuning_args.lora_target if model_args.use_kt: new_list = [] for m in target_modules: if m in ("down_proj", "up_proj", "gate_proj"): new_list.extend([f"mlp.{m}", f"shared_experts.{m}"]) elif m not in ("generate_linear", "orig_module", "prefill_linear"): new_list.append(m) target_modules[:] = new_list if finetuning_args.use_llama_pro: target_modules = find_expanded_modules(model, target_modules, finetuning_args.freeze_trainable_layers) target_modules = patch_target_modules(model, finetuning_args, target_modules) if ( finetuning_args.use_dora and getattr(model, "quantization_method", None) is not None and getattr(model, "quantization_method", None) != QuantizationMethod.BNB ): raise ValueError("DoRA is not compatible with PTQ-quantized models.") if model_args.resize_vocab and finetuning_args.additional_target is None: input_embeddings = model.get_input_embeddings() output_embeddings = model.get_output_embeddings() module_names = set() for name, module in model.named_modules(): if module in [input_embeddings, output_embeddings]: module_names.add(name.split(".")[-1]) finetuning_args.additional_target = module_names logger.warning_rank0("Vocab has been resized, add {} to trainable params.".format(",".join(module_names))) if finetuning_args.finetuning_type == "lora": peft_kwargs = { "r": finetuning_args.lora_rank, "target_modules": target_modules, "lora_alpha": finetuning_args.lora_alpha, "lora_dropout": finetuning_args.lora_dropout, "use_rslora": finetuning_args.use_rslora, "use_dora": finetuning_args.use_dora, "modules_to_save": finetuning_args.additional_target, } elif finetuning_args.finetuning_type == "oft": peft_kwargs = { "r": finetuning_args.oft_rank, "oft_block_size": finetuning_args.oft_block_size, "target_modules": target_modules, "module_dropout": finetuning_args.module_dropout, "modules_to_save": finetuning_args.additional_target, } if model_args.use_kt: if finetuning_args.finetuning_type == "oft": raise ValueError("KTransformers is currently not supported for OFT.") if finetuning_args.finetuning_type == "lora": peft_config = LoraConfig( task_type=TaskType.CAUSAL_LM, inference_mode=False, **peft_kwargs, ) else: raise ValueError("KTransformers is currently only supported for LoRA.") model = get_kt_peft_model(model, peft_config) print(f"KT_model:{model}") elif model_args.use_unsloth: if finetuning_args.finetuning_type == "oft": raise ValueError("Unsloth is currently not supported for OFT.") model = get_unsloth_peft_model(model, model_args, peft_kwargs) else: if finetuning_args.pissa_init: if finetuning_args.pissa_iter == -1: logger.info_rank0("Using PiSSA initialization.") peft_kwargs["init_lora_weights"] = "pissa" else: logger.info_rank0(f"Using PiSSA initialization with FSVD steps {finetuning_args.pissa_iter}.") peft_kwargs["init_lora_weights"] = f"pissa_niter_{finetuning_args.pissa_iter}" if finetuning_args.finetuning_type == "lora": peft_config = LoraConfig( task_type=TaskType.CAUSAL_LM, inference_mode=False, **peft_kwargs, ) elif finetuning_args.finetuning_type == "oft": peft_config = OFTConfig( task_type=TaskType.CAUSAL_LM, inference_mode=False, **peft_kwargs, ) model = get_peft_model(model, peft_config) if is_trainable and cast_trainable_params_to_fp32: for param in filter(lambda p: p.requires_grad, model.parameters()): param.data = param.data.to(torch.float32) return model def init_adapter( config: "PretrainedConfig", model: "PreTrainedModel", model_args: "ModelArguments", finetuning_args: "FinetuningArguments", is_trainable: bool, ) -> "PreTrainedModel": r"""Initialize the adapters. Support full-parameter, freeze and LoRA training. Note that the trainable parameters must be cast to float32. """ if is_trainable and getattr(model, "quantization_method", None) is not None: if finetuning_args.finetuning_type not in ["lora", "oft"]: raise ValueError("Quantized models can only be used for the LoRA or OFT tuning.") if finetuning_args.pissa_init: raise ValueError("Cannot initialize PiSSA adapter on quantized models.") # cast trainable parameters to float32 if: # 1. is_trainable and not pure_bf16 and not badam and quantization_bit is not None (qlora) # 2. is_trainable and not pure_bf16 and not badam and not zero3 (zero3 already in fp32) cast_trainable_params_to_fp32 = False if not is_trainable: pass elif finetuning_args.pure_bf16 or finetuning_args.use_badam: logger.info_rank0("Pure bf16 / BAdam detected, remaining trainable params in half precision.") elif model_args.quantization_bit is None and is_deepspeed_zero3_enabled(): logger.info_rank0("DeepSpeed ZeRO3 detected, remaining trainable params in float32.") else: logger.info_rank0("Upcasting trainable params to float32.") cast_trainable_params_to_fp32 = True if finetuning_args.finetuning_type == "full": _setup_full_tuning(model, finetuning_args, is_trainable, cast_trainable_params_to_fp32) elif finetuning_args.finetuning_type == "freeze": _setup_freeze_tuning(model, finetuning_args, is_trainable, cast_trainable_params_to_fp32) elif finetuning_args.finetuning_type in ["lora", "oft"]: model = _setup_lora_tuning( config, model, model_args, finetuning_args, is_trainable, cast_trainable_params_to_fp32 ) else: raise NotImplementedError(f"Unknown finetuning type: {finetuning_args.finetuning_type}.") return model
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/model/model_utils/kv_cache.py
src/llamafactory/model/model_utils/kv_cache.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...extras import logging logger = logging.get_logger(__name__) if TYPE_CHECKING: from transformers import PretrainedConfig from ...hparams import ModelArguments def configure_kv_cache(config: "PretrainedConfig", model_args: "ModelArguments", is_trainable: bool) -> None: if not is_trainable: setattr(config, "use_cache", model_args.use_kv_cache) if hasattr(config, "text_config"): setattr(config.text_config, "use_cache", model_args.use_kv_cache) if model_args.use_kv_cache: logger.info_rank0("KV cache is enabled for faster generation.") else: logger.info_rank0("KV cache is disabled.") else: setattr(config, "use_cache", False) if hasattr(config, "text_config"): setattr(config.text_config, "use_cache", False) logger.info_rank0("KV cache is disabled during training.")
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/model/model_utils/quantization.py
src/llamafactory/model/model_utils/quantization.py
# Copyright 2025 HuggingFace Inc. and the LlamaFactory team. # # This code is inspired by the HuggingFace's Transformers and Optimum library. # https://github.com/huggingface/transformers/blob/v4.41.0/src/transformers/utils/quantization_config.py # https://github.com/huggingface/optimum/blob/v1.20.0/optimum/gptq/data.py # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import os import random from typing import TYPE_CHECKING, Any import torch from datasets import load_dataset from transformers import BitsAndBytesConfig, EetqConfig, GPTQConfig, HqqConfig from transformers.integrations import is_deepspeed_zero3_enabled from transformers.modeling_utils import is_fsdp_enabled from ...extras import logging from ...extras.constants import FILEEXT2TYPE, QuantizationMethod from ...extras.misc import check_version, get_current_device if TYPE_CHECKING: from transformers import PretrainedConfig, PreTrainedTokenizer from ...hparams import ModelArguments logger = logging.get_logger(__name__) def _get_quantization_dataset(tokenizer: "PreTrainedTokenizer", model_args: "ModelArguments") -> list[dict[str, Any]]: r"""Prepare the tokenized dataset to perform AutoGPTQ. Do not use tensor output for JSON serialization.""" if os.path.isfile(model_args.export_quantization_dataset): data_path = FILEEXT2TYPE.get(model_args.export_quantization_dataset.split(".")[-1], None) data_files = model_args.export_quantization_dataset else: data_path = model_args.export_quantization_dataset data_files = None dataset = load_dataset( path=data_path, data_files=data_files, split="train", cache_dir=model_args.cache_dir, token=model_args.hf_hub_token, ) samples = [] maxlen = model_args.export_quantization_maxlen for _ in range(model_args.export_quantization_nsamples): n_try = 0 while True: if n_try > 100: raise ValueError("Cannot find satisfying example, considering decrease `export_quantization_maxlen`.") sample_idx = random.randint(0, len(dataset) - 1) sample: dict[str, torch.Tensor] = tokenizer(dataset[sample_idx]["text"], return_tensors="pt") n_try += 1 if sample["input_ids"].size(1) > maxlen: break # TODO: fix large maxlen word_idx = random.randint(0, sample["input_ids"].size(1) - maxlen - 1) input_ids = sample["input_ids"][:, word_idx : word_idx + maxlen] attention_mask = sample["attention_mask"][:, word_idx : word_idx + maxlen] samples.append({"input_ids": input_ids.tolist(), "attention_mask": attention_mask.tolist()}) return samples def configure_quantization( config: "PretrainedConfig", tokenizer: "PreTrainedTokenizer", model_args: "ModelArguments", is_trainable: bool, init_kwargs: dict[str, Any], ) -> None: r"""Priority: PTQ-quantized (train/infer) > AutoGPTQ (export) > On-the-fly quantization (train/infer).""" if getattr(config, "quantization_config", None): # ptq if model_args.quantization_bit is not None: logger.warning_rank0("`quantization_bit` will not affect on the PTQ-quantized models.") quantization_config: dict[str, Any] = getattr(config, "quantization_config", None) quant_method = quantization_config.get("quant_method", "") if quant_method not in (QuantizationMethod.MXFP4, QuantizationMethod.FP8) and ( is_deepspeed_zero3_enabled() or is_fsdp_enabled() ): # mxfp4 will dequant the model weights raise ValueError("DeepSpeed ZeRO-3 or FSDP is incompatible with PTQ-quantized models.") if quant_method == QuantizationMethod.MXFP4: from transformers import Mxfp4Config quant_config = Mxfp4Config(dequantize=True) init_kwargs["quantization_config"] = quant_config init_kwargs["ignore_mismatched_sizes"] = True if quant_method == QuantizationMethod.FP8: from transformers import FineGrainedFP8Config quant_config = FineGrainedFP8Config(dequantize=True) init_kwargs["quantization_config"] = quant_config init_kwargs["ignore_mismatched_sizes"] = True if quant_method == QuantizationMethod.GPTQ: check_version("gptqmodel>=2.0.0", mandatory=True) quantization_config.pop("disable_exllama", None) # remove deprecated args quantization_config["use_exllama"] = False # disable exllama if quant_method == QuantizationMethod.AWQ: check_version("autoawq", mandatory=True) if quant_method == QuantizationMethod.AQLM: check_version("aqlm>=1.1.0", mandatory=True) quantization_config["bits"] = 2 quant_bits = quantization_config.get("bits", "?") logger.info_rank0(f"Loading {quant_bits}-bit {quant_method.upper()}-quantized model.") elif model_args.export_quantization_bit is not None: # gptqmodel if model_args.export_quantization_bit not in [8, 4, 3, 2]: raise ValueError("AutoGPTQ only accepts 2/3/4/8-bit quantization.") check_version("optimum>=1.24.0", mandatory=True) check_version("gptqmodel>=2.0.0", mandatory=True) from accelerate.utils import get_max_memory if getattr(config, "model_type", None) == "chatglm": raise ValueError("ChatGLM model is not supported yet.") try: from optimum.gptq import utils as gq_utils if "language_model.model.layers" not in gq_utils.BLOCK_PATTERNS: gq_utils.BLOCK_PATTERNS.insert(0, "language_model.model.layers") except ImportError: pass block_name_to_quantize = None if getattr(config, "model_type", None) in ["gemma3", "paligemma"]: block_name_to_quantize = "language_model.model.layers" init_kwargs["quantization_config"] = GPTQConfig( bits=model_args.export_quantization_bit, tokenizer=tokenizer, dataset=_get_quantization_dataset(tokenizer, model_args), block_name_to_quantize=block_name_to_quantize, ) init_kwargs["device_map"] = "auto" init_kwargs["max_memory"] = get_max_memory() model_args.compute_dtype = torch.float16 # force fp16 for gptqmodel logger.info_rank0(f"Quantizing model to {model_args.export_quantization_bit} bit with GPTQModel.") elif model_args.quantization_bit is not None: # on-the-fly if model_args.quantization_method == QuantizationMethod.BNB: if model_args.quantization_bit == 8: check_version("bitsandbytes>=0.37.0", mandatory=True) init_kwargs["quantization_config"] = BitsAndBytesConfig(load_in_8bit=True) elif model_args.quantization_bit == 4: check_version("bitsandbytes>=0.39.0", mandatory=True) init_kwargs["quantization_config"] = BitsAndBytesConfig( load_in_4bit=True, bnb_4bit_compute_dtype=model_args.compute_dtype, bnb_4bit_use_double_quant=model_args.double_quantization, bnb_4bit_quant_type=model_args.quantization_type, bnb_4bit_quant_storage=model_args.compute_dtype, # crucial for fsdp+qlora ) else: raise ValueError("Bitsandbytes only accepts 4-bit or 8-bit quantization.") # Do not assign device map if: # 1. deepspeed zero3 or fsdp (train) # 2. auto quantization device map (inference) if is_deepspeed_zero3_enabled() or is_fsdp_enabled() or model_args.quantization_device_map == "auto": if model_args.quantization_bit != 4: raise ValueError("Only 4-bit quantized model can use fsdp+qlora or auto device map.") check_version("bitsandbytes>=0.43.0", mandatory=True) else: init_kwargs["device_map"] = {"": get_current_device()} # change auto device map for inference logger.info_rank0(f"Quantizing model to {model_args.quantization_bit} bit with bitsandbytes.") elif model_args.quantization_method == QuantizationMethod.HQQ: if model_args.quantization_bit not in [8, 6, 5, 4, 3, 2, 1]: raise ValueError("HQQ only accepts 1/2/3/4/5/6/8-bit quantization.") if is_deepspeed_zero3_enabled() or is_fsdp_enabled(): raise ValueError("HQQ quantization is incompatible with DeepSpeed ZeRO-3 or FSDP.") check_version("hqq", mandatory=True) init_kwargs["quantization_config"] = HqqConfig( nbits=model_args.quantization_bit, quant_zero=False, quant_scale=False, axis=0 ) # use ATEN kernel (axis=0) for performance logger.info_rank0(f"Quantizing model to {model_args.quantization_bit} bit with HQQ.") elif model_args.quantization_method == QuantizationMethod.EETQ: if model_args.quantization_bit != 8: raise ValueError("EETQ only accepts 8-bit quantization.") if is_deepspeed_zero3_enabled() or is_fsdp_enabled(): raise ValueError("EETQ quantization is incompatible with DeepSpeed ZeRO-3 or FSDP.") check_version("eetq", mandatory=True) init_kwargs["quantization_config"] = EetqConfig() logger.info_rank0(f"Quantizing model to {model_args.quantization_bit} bit with EETQ.")
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/model/model_utils/liger_kernel.py
src/llamafactory/model/model_utils/liger_kernel.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import inspect from typing import TYPE_CHECKING from ...extras import logging if TYPE_CHECKING: from transformers import PretrainedConfig from ...hparams import ModelArguments logger = logging.get_logger(__name__) def apply_liger_kernel( config: "PretrainedConfig", model_args: "ModelArguments", is_trainable: bool, require_logits: bool, ) -> None: if not is_trainable or not model_args.enable_liger_kernel: return model_type = getattr(config, "model_type", None) if model_type == "gemma": from liger_kernel.transformers import apply_liger_kernel_to_gemma as apply_liger_kernel elif model_type == "gemma2": from liger_kernel.transformers import apply_liger_kernel_to_gemma2 as apply_liger_kernel elif model_type == "gemma3": from liger_kernel.transformers import apply_liger_kernel_to_gemma3 as apply_liger_kernel elif model_type == "gemma3_text": from liger_kernel.transformers import apply_liger_kernel_to_gemma3_text as apply_liger_kernel elif model_type == "glm4": from liger_kernel.transformers import apply_liger_kernel_to_glm4 as apply_liger_kernel elif model_type == "glm4v": from liger_kernel.transformers import apply_liger_kernel_to_glm4v as apply_liger_kernel elif model_type == "granite": from liger_kernel.transformers import apply_liger_kernel_to_granite as apply_liger_kernel elif model_type == "llama": from liger_kernel.transformers import apply_liger_kernel_to_llama as apply_liger_kernel elif model_type == "llava": from liger_kernel.transformers import apply_liger_kernel_to_llava as apply_liger_kernel elif model_type == "mistral": from liger_kernel.transformers import apply_liger_kernel_to_mistral as apply_liger_kernel elif model_type == "mixtral": from liger_kernel.transformers import apply_liger_kernel_to_mixtral as apply_liger_kernel elif model_type == "mllama": from liger_kernel.transformers import apply_liger_kernel_to_mllama as apply_liger_kernel elif model_type == "olmo2": from liger_kernel.transformers import apply_liger_kernel_to_olmo2 as apply_liger_kernel elif model_type == "paligemma": from liger_kernel.transformers import apply_liger_kernel_to_paligemma as apply_liger_kernel elif model_type == "phi3": from liger_kernel.transformers import apply_liger_kernel_to_phi3 as apply_liger_kernel elif model_type == "qwen2": from liger_kernel.transformers import apply_liger_kernel_to_qwen2 as apply_liger_kernel elif model_type == "qwen2_vl": from liger_kernel.transformers import apply_liger_kernel_to_qwen2_vl as apply_liger_kernel elif model_type == "qwen2_5_vl": from liger_kernel.transformers import apply_liger_kernel_to_qwen2_5_vl as apply_liger_kernel elif model_type == "qwen3": from liger_kernel.transformers import apply_liger_kernel_to_qwen3 as apply_liger_kernel elif model_type == "qwen3_moe": from liger_kernel.transformers import apply_liger_kernel_to_qwen3_moe as apply_liger_kernel elif model_type == "gpt_oss": try: from liger_kernel.transformers import apply_liger_kernel_to_gpt_oss as apply_liger_kernel except ImportError: logger.warning_rank0("Please install liger-kernel from https://github.com/Comet0322/Liger-Kernel.") return else: logger.warning_rank0("Current model does not support liger kernel.") return if require_logits and "fused_linear_cross_entropy" in inspect.signature(apply_liger_kernel).parameters: logger.info_rank0("Current training stage does not support chunked cross entropy.") kwargs = {"fused_linear_cross_entropy": False, "cross_entropy": True} else: kwargs = {} apply_liger_kernel(**kwargs) logger.info_rank0("Liger kernel has been applied to the model.")
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/model/model_utils/packing.py
src/llamafactory/model/model_utils/packing.py
# Copyright 2025 Musab Gultekin and the LlamaFactory team. # # This code is based on the Musab Gultekin's functionary library. # https://github.com/MeetKai/functionary/blob/main/functionary/train/packing/monkey_patch_packing.py # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # # MIT License # # Copyright (c) 2023 Musab Gultekin # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitation the rights # to use, copy, modify, merge, publish, distribute, sublicense, and/or sell # copies of the Software, and to permit persons to whom the Software is # furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in all # copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE # SOFTWARE. from typing import TYPE_CHECKING import torch import torch.nn.functional as F from ...extras import logging if TYPE_CHECKING: from ...hparams import ModelArguments logger = logging.get_logger(__name__) def get_seqlens_in_batch(attention_mask: "torch.Tensor") -> "torch.Tensor": r"""Get the sequence lengths in the current batch. e.g. ```python # input [ [1, 1, 2, 2, 2, 0], [1, 2, 2, 3, 3, 3], ] # output [2, 3, 1, 2, 3] ``` """ bsz = attention_mask.size(0) dtype, device = attention_mask.dtype, attention_mask.device max_num = torch.max(attention_mask).item() counts: torch.Tensor = torch.zeros((bsz, max_num), dtype=dtype, device=device) for i in range(max_num): counts[:, i] = torch.sum(attention_mask == (i + 1), dim=-1) counts = counts.flatten() seqlens = counts[counts.nonzero().squeeze(dim=-1)] return seqlens def get_unpad_data(attention_mask: "torch.Tensor") -> tuple["torch.Tensor", "torch.Tensor", int]: r"""Prepare the indices and seqlens for flash attn varlen function. Returns: indices: indices of non-masked tokens from the flattened sequence. cu_seqlens: the cumulative sequence lengths in the current batch, always starts from 0. max_seqlen_in_batch: the largest seqlen in the current batch. e.g. ```python # input [ [1, 1, 2, 2, 2, 0], [1, 2, 2, 3, 3, 3], ] # output [0, 1, 2, 3, 4, 6, 7, 8, 9, 10, 11] [0, 2, 5, 6, 8, 11] 3 ``` """ seqlens_in_batch = get_seqlens_in_batch(attention_mask) indices = torch.nonzero(attention_mask.flatten(), as_tuple=False).flatten() max_seqlen_in_batch = seqlens_in_batch.max().item() cu_seqlens = F.pad(torch.cumsum(seqlens_in_batch, dim=0, dtype=torch.int32), (1, 0)) return indices, cu_seqlens, max_seqlen_in_batch def configure_packing(model_args: "ModelArguments", is_trainable: bool) -> None: if not is_trainable or not model_args.block_diag_attn: return import transformers.modeling_flash_attention_utils transformers.modeling_flash_attention_utils._get_unpad_data = get_unpad_data logger.info_rank0("Using block diagonal attention for sequence packing without cross-attention.")
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/model/model_utils/checkpointing.py
src/llamafactory/model/model_utils/checkpointing.py
# Copyright 2025 HuggingFace Inc., Daniel Han-Chen & the Unsloth team and the LlamaFactory team. # # This code is inspired by the HuggingFace's Transformers and PEFT library, # https://github.com/huggingface/transformers/blob/v4.40.0/src/transformers/modeling_utils.py # https://github.com/huggingface/peft/blob/v0.10.0/src/peft/utils/other.py # and the Unsloth library. # https://github.com/unslothai/unsloth/blob/July-2024/unsloth/models/_utils.py # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import inspect import os from collections.abc import Callable from functools import WRAPPER_ASSIGNMENTS, partial, wraps from types import MethodType from typing import TYPE_CHECKING, Any, Optional, Union import torch from ...extras import logging from ...extras.constants import LAYERNORM_NAMES if TYPE_CHECKING: from transformers import PreTrainedModel from ...hparams import ModelArguments logger = logging.get_logger(__name__) def get_unsloth_gradient_checkpointing_func() -> Callable: class UnslothGradientCheckpointing(torch.autograd.Function): r"""Saves VRAM by smartly offloading to RAM.""" @staticmethod @torch.cuda.amp.custom_fwd def forward( ctx: "torch.autograd.Function", forward_function: "torch.Module", hidden_states: "torch.Tensor", *args: Union["torch.Tensor", Any], ) -> "torch.Tensor": saved_hidden_states = hidden_states.to("cpu", non_blocking=True) with torch.no_grad(): outputs = forward_function(hidden_states, *args) ctx.save_for_backward(saved_hidden_states) ctx.forward_function = forward_function ctx.args = args return outputs @staticmethod @torch.cuda.amp.custom_bwd def backward(ctx: "torch.autograd.Function", grad_output: "torch.Tensor") -> "torch.Tensor": (hidden_states,) = ctx.saved_tensors hidden_states = hidden_states.to("cuda", non_blocking=True).detach() hidden_states.requires_grad_(True) with torch.enable_grad(): outputs = ctx.forward_function(hidden_states, *ctx.args) output = outputs[0] if isinstance(outputs, tuple) else outputs torch.autograd.backward(output, grad_output) return (None, hidden_states.grad) + (None,) * len(ctx.args) return UnslothGradientCheckpointing.apply def get_custom_gradient_checkpointing_func(gradient_checkpointing_func: Callable) -> Callable: r"""Only applies gradient checkpointing to trainable layers.""" @wraps(gradient_checkpointing_func, assigned=WRAPPER_ASSIGNMENTS + ("__self__",)) def custom_gradient_checkpointing_func(func: Callable, *args: Union["torch.Tensor", Any], **kwargs): if isinstance(func, partial): module: torch.nn.Module = func.func.__self__ else: module: torch.nn.Module = func.__self__ has_grad = False if any(param.requires_grad for param in module.parameters()): has_grad = True for arg in args: if torch.is_tensor(arg) and torch.is_floating_point(arg): arg.requires_grad_(True) break # assume the first tensor is always the hidden states if has_grad: return gradient_checkpointing_func(func, *args, **kwargs) else: return func(*args, **kwargs) return custom_gradient_checkpointing_func def _gradient_checkpointing_enable( self: "PreTrainedModel", gradient_checkpointing_kwargs: Optional[dict[str, Any]] = None, use_unsloth_gc: bool = False, ) -> None: r"""Activates gradient checkpointing for the current model. Modification of the original method to enable gradient checkpointing for block-wise optimizer. """ from torch.utils.checkpoint import checkpoint if not self.supports_gradient_checkpointing: raise ValueError(f"{self.__class__.__name__} does not support gradient checkpointing.") if gradient_checkpointing_kwargs is None: gradient_checkpointing_kwargs = {"use_reentrant": True} if use_unsloth_gc: gradient_checkpointing_func = get_unsloth_gradient_checkpointing_func() else: gradient_checkpointing_func = partial(checkpoint, **gradient_checkpointing_kwargs) gradient_checkpointing_func = get_custom_gradient_checkpointing_func(gradient_checkpointing_func) if "value" in inspect.signature(self._set_gradient_checkpointing).parameters: # old GC format self.apply(partial(self._set_gradient_checkpointing, value=True)) self.enable_input_require_grads() logger.warning_rank0_once("You are using the old GC format, some features (e.g. BAdam) will be invalid.") else: # have already enabled input require gradients self._set_gradient_checkpointing(enable=True, gradient_checkpointing_func=gradient_checkpointing_func) def _fp32_forward_post_hook( module: "torch.nn.Module", args: tuple["torch.Tensor"], output: "torch.Tensor" ) -> "torch.Tensor": return output.to(torch.float32) def prepare_model_for_training(model: "PreTrainedModel", model_args: "ModelArguments") -> None: r"""Prepare the model before training. Include: (1) cast the layernorm in fp32 (2) make output embedding layer require grads (3) add the upcasting of the lm_head in fp32. """ if model_args.upcast_layernorm: logger.info_rank0("Upcasting layernorm weights in float32.") for name, param in model.named_parameters(): if param.ndim == 1 and any(ln_name in name for ln_name in LAYERNORM_NAMES): param.data = param.data.to(torch.float32) if ( os.environ.get("ACCELERATE_USE_FSDP", "false").lower() == "true" and int(os.environ.get("FSDP_VERSION", "1")) == 2 ): model_args.use_reentrant_gc = False logger.warning_rank0("You are using fsdp2, `use_reentrant_gc` has been set to False.") if not model_args.disable_gradient_checkpointing: if not getattr(model, "supports_gradient_checkpointing", False): logger.warning_rank0("Current model does not support gradient checkpointing.") else: # use_reentrant=False might increase VRAM usage (have not been empirically verified yet) # According to: https://github.com/huggingface/transformers/issues/28339 gradient_checkpointing_enable = partial( _gradient_checkpointing_enable, use_unsloth_gc=model_args.use_unsloth_gc ) model.gradient_checkpointing_enable = MethodType(gradient_checkpointing_enable, model) model.gradient_checkpointing_enable( gradient_checkpointing_kwargs={"use_reentrant": model_args.use_reentrant_gc} ) setattr(model.config, "use_cache", False) # turn off when gradient checkpointing is enabled logger.info_rank0("Gradient checkpointing enabled.") if model_args.upcast_lmhead_output: output_layer = model.get_output_embeddings() if isinstance(output_layer, torch.nn.Linear) and output_layer.weight.dtype != torch.float32: logger.info_rank0("Upcasting lm_head outputs in float32.") output_layer.register_forward_hook(_fp32_forward_post_hook)
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/model/model_utils/unsloth.py
src/llamafactory/model/model_utils/unsloth.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING, Any, Optional from ...extras import logging from ...extras.misc import get_current_device if TYPE_CHECKING: from transformers import PretrainedConfig, PreTrainedModel from ...hparams import FinetuningArguments, ModelArguments logger = logging.get_logger(__name__) def _get_unsloth_kwargs( config: "PretrainedConfig", model_name_or_path: str, model_args: "ModelArguments", finetuning_args: "FinetuningArguments", ) -> dict[str, Any]: return { "model_name": model_name_or_path, "max_seq_length": model_args.model_max_length or 4096, "dtype": model_args.compute_dtype, "load_in_4bit": model_args.quantization_bit == 4, "token": model_args.hf_hub_token, "full_finetuning": finetuning_args.finetuning_type == "full", "device_map": {"": get_current_device()}, "rope_scaling": getattr(config, "rope_scaling", None), "fix_tokenizer": False, "trust_remote_code": model_args.trust_remote_code, "use_gradient_checkpointing": "unsloth", } def load_unsloth_pretrained_model( config: "PretrainedConfig", model_args: "ModelArguments", finetuning_args: "FinetuningArguments" ) -> Optional["PreTrainedModel"]: r"""Optionally load pretrained model with unsloth. Used in training.""" from unsloth import FastLanguageModel # type: ignore unsloth_kwargs = _get_unsloth_kwargs(config, model_args.model_name_or_path, model_args, finetuning_args) try: model, _ = FastLanguageModel.from_pretrained(**unsloth_kwargs) except NotImplementedError: logger.warning_rank0("Unsloth does not support model type {}.".format(getattr(config, "model_type", None))) model = None model_args.use_unsloth = False return model def get_unsloth_peft_model( model: "PreTrainedModel", model_args: "ModelArguments", peft_kwargs: dict[str, Any] ) -> "PreTrainedModel": r"""Get the peft model for the pretrained model with unsloth. Used in training.""" from unsloth import FastLanguageModel # type: ignore unsloth_peft_kwargs = { "model": model, "max_seq_length": model_args.model_max_length, "use_gradient_checkpointing": "unsloth", } return FastLanguageModel.get_peft_model(**peft_kwargs, **unsloth_peft_kwargs) def load_unsloth_peft_model( config: "PretrainedConfig", model_args: "ModelArguments", finetuning_args: "FinetuningArguments", is_trainable: bool, ) -> "PreTrainedModel": r"""Load peft model with unsloth. Used in both training and inference.""" from unsloth import FastLanguageModel # type: ignore unsloth_kwargs = _get_unsloth_kwargs(config, model_args.adapter_name_or_path[0], model_args, finetuning_args) try: if not is_trainable: unsloth_kwargs["use_gradient_checkpointing"] = False model, _ = FastLanguageModel.from_pretrained(**unsloth_kwargs) except NotImplementedError: raise ValueError("Unsloth does not support model type {}.".format(getattr(config, "model_type", None))) if not is_trainable: FastLanguageModel.for_inference(model) return model
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/model/model_utils/visual.py
src/llamafactory/model/model_utils/visual.py
# Copyright 2025 HuggingFace Inc. and the LlamaFactory team. # # This code is inspired by the HuggingFace's Transformers library. # https://github.com/huggingface/transformers/blob/v4.40.0/src/transformers/models/llava/modeling_llava.py # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from dataclasses import dataclass from typing import TYPE_CHECKING, Optional import torch import transformers import transformers.models from transformers.activations import ACT2FN from ...extras import logging from ...extras.packages import is_transformers_version_greater_than if TYPE_CHECKING: from transformers import LlavaConfig, PretrainedConfig, PreTrainedModel from ...hparams import FinetuningArguments, ModelArguments logger = logging.get_logger(__name__) transformers_logger = transformers.utils.logging.get_logger(__name__) @dataclass class CompositeModel: model_type: str projector_key: str vision_model_keys: list[str] language_model_keys: list[str] lora_conflict_keys: list[str] def get_projector(self, module: "torch.nn.Module") -> "torch.nn.Module": for key in self.projector_key.split("."): module = getattr(module, key) return module COMPOSITE_MODELS: dict[str, "CompositeModel"] = {} def _register_composite_model( model_type: str, projector_key: Optional[str] = None, vision_model_keys: Optional[list[str]] = None, language_model_keys: Optional[list[str]] = None, lora_conflict_keys: Optional[list[str]] = None, ): r"""Register a new composite model. Args: model_type: model type projector_key: multi_modal_projector vision_model_keys: vision_tower language_model_keys: language_model lora_conflict_keys: None """ COMPOSITE_MODELS[model_type] = CompositeModel( model_type=model_type, projector_key=projector_key or "multi_modal_projector", vision_model_keys=vision_model_keys or ["vision_tower"], language_model_keys=language_model_keys or ["language_model", "lm_head"], lora_conflict_keys=lora_conflict_keys or [], ) class LlavaMultiModalProjectorForYiVL(torch.nn.Module): def __init__(self, config: "LlavaConfig") -> None: super().__init__() self.config = config if config is None: return self.linear_1 = torch.nn.Linear(config.vision_config.hidden_size, config.text_config.hidden_size, bias=True) self.linear_2 = torch.nn.LayerNorm(config.text_config.hidden_size, bias=True) self.linear_3 = torch.nn.Linear(config.text_config.hidden_size, config.text_config.hidden_size, bias=True) self.linear_4 = torch.nn.LayerNorm(config.text_config.hidden_size, bias=True) self.act = ACT2FN[config.projector_hidden_act] def forward(self, image_features: "torch.Tensor") -> "torch.Tensor": hidden_states = self.linear_1(image_features) hidden_states = self.linear_2(hidden_states) hidden_states = self.act(hidden_states) hidden_states = self.linear_3(hidden_states) hidden_states = self.linear_4(hidden_states) if hidden_states.dtype == torch.float32: if torch.is_autocast_enabled(): target_dtype = torch.get_autocast_gpu_dtype() elif hasattr(self.config, "_pre_quantization_dtype"): target_dtype = self.config._pre_quantization_dtype else: target_dtype = self.linear_1.weight.dtype transformers_logger.warning_once("The hidden states seems to be silently casted in float32.") hidden_states = hidden_states.to(target_dtype) return hidden_states class LlavaMultiModalProjectorForYiVLForVLLM(LlavaMultiModalProjectorForYiVL): def __init__(self, vision_hidden_size: int, text_hidden_size: int, projector_hidden_act: str) -> None: super().__init__(config=None) self.linear_1 = torch.nn.Linear(vision_hidden_size, text_hidden_size, bias=True) self.linear_2 = torch.nn.LayerNorm(text_hidden_size, bias=True) self.linear_3 = torch.nn.Linear(text_hidden_size, text_hidden_size, bias=True) self.linear_4 = torch.nn.LayerNorm(text_hidden_size, bias=True) self.act = ACT2FN[projector_hidden_act] def autocast_projector_dtype(model: "PreTrainedModel", model_args: "ModelArguments") -> None: r"""Cast projector output to half precision for fine-tuning quantized VLMs.""" def _mm_projector_forward_post_hook( module: "torch.nn.Module", args: tuple["torch.Tensor"], output: "torch.Tensor" ) -> "torch.Tensor": return output.to(model_args.compute_dtype) if getattr(model, "quantization_method", None): model_type = getattr(model.config, "model_type", None) if model_type in COMPOSITE_MODELS: mm_projector = COMPOSITE_MODELS[model_type].get_projector(model) else: return logger.info_rank0(f"Casting multimodal projector outputs in {model_args.compute_dtype}.") mm_projector.register_forward_hook(_mm_projector_forward_post_hook) def configure_visual_model(config: "PretrainedConfig") -> None: r"""Patch VLMs before loading them.""" if getattr(config, "text_config", None) and not getattr(config, "hidden_size", None): # required for ds zero3 and valuehead models setattr(config, "hidden_size", getattr(config.text_config, "hidden_size", None)) if getattr(config, "is_yi_vl_derived_model", None): logger.info_rank0("Detected Yi-VL model, applying projector patch.") transformers.models.llava.modeling_llava.LlavaMultiModalProjector = LlavaMultiModalProjectorForYiVL def get_forbidden_modules(config: "PretrainedConfig", finetuning_args: "FinetuningArguments") -> set[str]: r"""Freeze vision tower and language model for VLM full/freeze tuning.""" model_type = getattr(config, "model_type", None) forbidden_modules = set() if model_type in COMPOSITE_MODELS: if finetuning_args.freeze_vision_tower: vision_model_keys = COMPOSITE_MODELS[model_type].vision_model_keys logger.info_rank0(f"Set vision model not trainable: {vision_model_keys}.") forbidden_modules.update(vision_model_keys) if finetuning_args.freeze_multi_modal_projector: projector_key = COMPOSITE_MODELS[model_type].projector_key logger.info_rank0(f"Set multi model projector not trainable: {projector_key}.") forbidden_modules.add(projector_key) if finetuning_args.freeze_language_model: language_model_keys = COMPOSITE_MODELS[model_type].language_model_keys logger.info_rank0(f"Set language model not trainable: {language_model_keys}.") forbidden_modules.update(language_model_keys) return forbidden_modules def patch_target_modules( model: "PreTrainedModel", finetuning_args: "FinetuningArguments", target_modules: list[str] ) -> list[str]: r"""Freeze vision tower for VLM LoRA tuning.""" model_type = getattr(model.config, "model_type", None) if model_type in COMPOSITE_MODELS: forbidden_modules = get_forbidden_modules(model.config, finetuning_args) forbidden_modules.update(COMPOSITE_MODELS[model_type].lora_conflict_keys) module_names = [] for name, _ in model.named_modules(): if any(target_module in name for target_module in target_modules) and not any( forbidden_module in name for forbidden_module in forbidden_modules ): module_names.append(name) return module_names else: return target_modules _register_composite_model( model_type="dots_ocr", projector_key="vision_tower.merger", vision_model_keys=["vision_tower"], language_model_keys=["model", "lm_head"], lora_conflict_keys=["merger"], ) _register_composite_model( model_type="gemma3", ) _register_composite_model( model_type="gemma3n", vision_model_keys=["vision_tower", "audio_tower"], lora_conflict_keys=["timm_model", "subsample_conv_projection"], ) # copied from qwen2vl _register_composite_model( model_type="glm4v", projector_key="visual.merger", vision_model_keys=["visual.patch_embed", "visual.blocks"], language_model_keys=["language_model", "lm_head"], lora_conflict_keys=["patch_embed"], ) _register_composite_model( model_type="glm4v_moe", projector_key="visual.merger", vision_model_keys=["visual.patch_embed", "visual.blocks"], language_model_keys=["language_model", "lm_head"], lora_conflict_keys=["patch_embed"], ) _register_composite_model( model_type="internvl", ) _register_composite_model( model_type="interns1", ) _register_composite_model( model_type="Keye", projector_key="mlp_AR", vision_model_keys=["visual.vision_model.patch_embedding", "visual.vision_model.encoder"], language_model_keys=["model", "lm_head"], lora_conflict_keys=["patch_embedding"], ) _register_composite_model( model_type="kimi_vl", ) _register_composite_model( model_type="llama4", vision_model_keys=["vision_model"], ) _register_composite_model( model_type="llava", ) _register_composite_model( model_type="llava_next", ) _register_composite_model( model_type="llava_next_video", ) _register_composite_model( model_type="minicpmv", projector_key="resampler", vision_model_keys=["vpm"], language_model_keys=["llm"], ) _register_composite_model( model_type="minicpmo", projector_key="resampler", vision_model_keys=["vpm", "apm", "audio_avg_pooler", "audio_projection_layer", "tts"], language_model_keys=["llm"], lora_conflict_keys=["audio_projection_layer"], ) _register_composite_model( model_type="mistral3", projector_key="model.multi_modal_projector", ) _register_composite_model( model_type="mllama", vision_model_keys=["vision_model"], ) _register_composite_model( model_type="paligemma", ) _register_composite_model( model_type="qwen2_audio", vision_model_keys=["audio_tower"], ) _register_composite_model( model_type="qwen2_5_omni_thinker", projector_key="visual.merger", vision_model_keys=["visual.patch_embed", "visual.blocks", "audio_tower"], language_model_keys=["model", "lm_head"], lora_conflict_keys=["patch_embed"], ) _register_composite_model( model_type="qwen2_vl", projector_key="visual.merger", vision_model_keys=["visual.patch_embed", "visual.blocks"], language_model_keys=["language_model", "lm_head"] if is_transformers_version_greater_than("4.52.0") else ["model", "lm_head"], lora_conflict_keys=["patch_embed"], ) _register_composite_model( model_type="qwen2_5_vl", projector_key="visual.merger", vision_model_keys=["visual.patch_embed", "visual.blocks"], language_model_keys=["language_model", "lm_head"] if is_transformers_version_greater_than("4.52.0") else ["model", "lm_head"], lora_conflict_keys=["patch_embed"], ) _register_composite_model( model_type="qwen3_vl", projector_key="visual.merger", vision_model_keys=["visual.patch_embed", "visual.blocks", "visual.deepstack_merger_list"], language_model_keys=["language_model", "lm_head"], lora_conflict_keys=["patch_embed"], ) _register_composite_model( model_type="qwen3_vl_moe", projector_key="visual.merger", vision_model_keys=["visual.patch_embed", "visual.blocks", "visual.deepstack_merger_list"], language_model_keys=["language_model", "lm_head"], lora_conflict_keys=["patch_embed"], ) _register_composite_model( model_type="qwen3_omni_moe_thinker", projector_key="visual.merger", vision_model_keys=["visual.patch_embed", "visual.blocks", "visual.deepstack_merger_list", "audio_tower"], language_model_keys=["model", "lm_head"], lora_conflict_keys=["patch_embed"], ) _register_composite_model( model_type="video_llava", )
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/model/model_utils/mod.py
src/llamafactory/model/model_utils/mod.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...extras.constants import MOD_SUPPORTED_MODELS if TYPE_CHECKING: from transformers import PretrainedConfig, PreTrainedModel from ...hparams import ModelArguments def load_mod_pretrained_model(**init_kwargs) -> "PreTrainedModel": from MoD import AutoMoDModelForCausalLM return AutoMoDModelForCausalLM.from_pretrained(**init_kwargs) def convert_pretrained_model_to_mod( model: "PreTrainedModel", config: "PretrainedConfig", model_args: "ModelArguments" ) -> "PreTrainedModel": from MoD import apply_mod_to_hf if getattr(config, "model_type", None) not in MOD_SUPPORTED_MODELS: raise ValueError("Current model is not supported by mixture-of-depth.") model = apply_mod_to_hf(model) model = model.to(model_args.compute_dtype) return model
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/model/model_utils/ktransformers.py
src/llamafactory/model/model_utils/ktransformers.py
# Copyright 2025 the KVCache.AI team, Approaching AI, and the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import importlib.util as _u from typing import TYPE_CHECKING, Any import torch from ...extras import logging from ...extras.misc import get_current_device if TYPE_CHECKING: from ...hparams import FinetuningArguments, ModelArguments from transformers import AutoConfig, AutoModelForCausalLM, PretrainedConfig, PreTrainedModel KT_AVAILABLE = _u.find_spec("ktransformers") is not None if KT_AVAILABLE: from ktransformers.models.modeling_deepseek import DeepseekV2ForCausalLM from ktransformers.models.modeling_deepseek_v3 import DeepseekV3ForCausalLM from ktransformers.models.modeling_llama import LlamaForCausalLM from ktransformers.models.modeling_mixtral import MixtralForCausalLM from ktransformers.models.modeling_qwen2_moe import Qwen2MoeForCausalLM from ktransformers.models.modeling_qwen3_moe import Qwen3MoeForCausalLM from ktransformers.optimize.optimize import optimize_and_load_gguf from ktransformers.server.config.config import Config from ktransformers.sft.lora import inject_lora_layer from ktransformers.util.custom_loader import GGUFLoader, SafeTensorLoader from ktransformers.util.globals import GLOBAL_CONFIG from ktransformers.util.utils import load_weights logger = logging.get_logger(__name__) def _get_kt_kwargs( config: "PretrainedConfig", model_name_or_path: str, model_args: "ModelArguments", finetuning_args: "FinetuningArguments", ) -> dict[str, Any]: return { "model_name": model_name_or_path, "max_seq_length": model_args.model_max_length or 4096, "dtype": model_args.compute_dtype, "load_in_4bit": model_args.quantization_bit == 4, "token": model_args.hf_hub_token, "full_finetuning": finetuning_args.finetuning_type == "full", "device_map": {"": get_current_device()}, "rope_scaling": getattr(config, "rope_scaling", None), "fix_tokenizer": False, "trust_remote_code": model_args.trust_remote_code, "use_gradient_checkpointing": "ktransformers", } def load_kt_pretrained_model(config: "PretrainedConfig", model_args: "ModelArguments") -> "PreTrainedModel": r"""Optionally load pretrained model with KTransformers. Used in training.""" custom_models = { "DeepseekV2ForCausalLM": DeepseekV2ForCausalLM, "DeepseekV3ForCausalLM": DeepseekV3ForCausalLM, "Qwen2MoeForCausalLM": Qwen2MoeForCausalLM, "Qwen3MoeForCausalLM": Qwen3MoeForCausalLM, "LlamaForCausalLM": LlamaForCausalLM, "MixtralForCausalLM": MixtralForCausalLM, } Config().cpu_infer = model_args.cpu_infer Config().chunk_size = model_args.chunk_size config = AutoConfig.from_pretrained(model_args.model_name_or_path, trust_remote_code=model_args.trust_remote_code) if model_args.mode == "long_context": assert config.architectures[0] == "LlamaForCausalLM", "only LlamaForCausalLM support long_context mode" torch.set_default_dtype(torch.float16) else: torch.set_default_dtype(config.torch_dtype) with torch.device("meta"): if config.architectures[0] in custom_models: print("using custom modeling_xxx.py.") if "Qwen2Moe" in config.architectures[0]: # Qwen2Moe must use flash_attention_2 to avoid overflow. config._attn_implementation = "flash_attention_2" if "Llama" in config.architectures[0]: config._attn_implementation = "eager" if "Mixtral" in config.architectures[0]: config._attn_implementation = "flash_attention_2" model = custom_models[config.architectures[0]](config) else: attn_implementation = "flash_attention_2" model = AutoModelForCausalLM.from_config( config, trust_remote_code=True, attn_implementation=attn_implementation ) optimize_config_path = model_args.kt_optimize_rule gguf_path = model_args.model_name_or_path assert optimize_config_path is not None, "optimize_config_path must be provided (path to YAML rules file)." assert gguf_path is not None, "gguf_path must be provided (path to a folder or .gguf file)." GLOBAL_CONFIG._config["mod"] = "infer" optimize_and_load_gguf(model, optimize_config_path, gguf_path, config) return model def get_kt_peft_model(model: "PreTrainedModel", peft_kwargs: dict[str, Any]) -> "PreTrainedModel": r"""Get the peft model for the pretrained model with KTransformers. Used in training.""" from ktransformers.sft.peft_utils.mapping import get_peft_model return get_peft_model(model, peft_kwargs) def load_kt_peft_model(model_args: "ModelArguments", model: "PreTrainedModel") -> "PreTrainedModel": r"""Load peft model with KTransformers. Used in both training and inference.""" load_adapter_name_or_path = model_args.adapter_name_or_path[0] if load_adapter_name_or_path.endswith(".gguf"): inject_lora_layer(model, load_adapter_name_or_path) adapter_gguf_loader = GGUFLoader(load_adapter_name_or_path) load_weights(model, adapter_gguf_loader, adapter_gguf=True) model.train() else: inject_lora_layer(model, load_adapter_name_or_path) adapter_loader = SafeTensorLoader(load_adapter_name_or_path) device = next(model.parameters()).device for key in adapter_loader.tensor_file_map.keys(): try: tensor = adapter_loader.load_tensor(key, device=device) model_key = key.replace("base_model.model.", "") model_key = model_key.replace(".weight", ".default.weight") model_key = model_key.replace(".default.default.weight", ".default.weight") param = model.get_parameter(model_key) param.data.copy_(tensor.data) print(f"Loaded adapter weight: {key} -> {model_key}") except AttributeError: print(f"Skipping {key}: not a model parameter") except KeyError: print(f"Key not found in model: {model_key} (original: {key})") return model
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/model/model_utils/misc.py
src/llamafactory/model/model_utils/misc.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...extras import logging from .visual import COMPOSITE_MODELS if TYPE_CHECKING: from transformers import PretrainedConfig, PreTrainedModel, PreTrainedTokenizer logger = logging.get_logger(__name__) def find_all_linear_modules(model: "PreTrainedModel", freeze_vision_tower: bool) -> list[str]: r"""Find all available modules to apply LoRA, GaLore or APOLLO.""" model_type = getattr(model.config, "model_type", None) forbidden_modules = {"lm_head"} if model_type == "chatglm": forbidden_modules.add("output_layer") elif model_type == "internlm2": forbidden_modules.add("output") if model_type in COMPOSITE_MODELS: forbidden_modules.add(COMPOSITE_MODELS[model_type].projector_key) if freeze_vision_tower and model_type in COMPOSITE_MODELS: forbidden_modules.update(COMPOSITE_MODELS[model_type].vision_model_keys) module_names = set() for name, module in model.named_modules(): if any(forbidden_module in name for forbidden_module in forbidden_modules): continue if "Linear" in module.__class__.__name__ and "Embedding" not in module.__class__.__name__: module_names.add(name.split(".")[-1]) logger.info_rank0("Found linear modules: {}".format(",".join(module_names))) return list(module_names) def find_expanded_modules(model: "PreTrainedModel", target_modules: list[str], num_layer_trainable: int) -> list[str]: r"""Find the modules in the expanded blocks to apply lora.""" num_layers = getattr(model.config, "num_hidden_layers", None) if not num_layers: raise ValueError("Model was not supported.") if num_layers % num_layer_trainable != 0: raise ValueError( f"`num_layers` {num_layers} should be divisible by `num_layer_trainable` {num_layer_trainable}." ) stride = num_layers // num_layer_trainable trainable_layer_ids = range(stride - 1, num_layers + stride - 1, stride) trainable_layers = [f".{idx:d}." for idx in trainable_layer_ids] module_names = [] for name, _ in model.named_modules(): if any(target_module in name for target_module in target_modules) and any( trainable_layer in name for trainable_layer in trainable_layers ): module_names.append(name) logger.info_rank0("Apply lora to layers: {}.".format(",".join(map(str, trainable_layer_ids)))) return module_names def register_autoclass(config: "PretrainedConfig", model: "PreTrainedModel", tokenizer: "PreTrainedTokenizer"): if "AutoConfig" in getattr(config, "auto_map", {}): config.__class__.register_for_auto_class() if "AutoModelForCausalLM" in getattr(config, "auto_map", {}): model.__class__.register_for_auto_class() if "AutoTokenizer" in tokenizer.init_kwargs.get("auto_map", {}): tokenizer.__class__.register_for_auto_class()
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/model/model_utils/moe.py
src/llamafactory/model/model_utils/moe.py
# Copyright 2025 the LlamaFactory team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING, Union import torch from torch import nn from torch.nn import functional as F from transformers.integrations import is_deepspeed_zero3_enabled from ...extras.misc import check_version from ...extras.packages import is_transformers_version_greater_than if TYPE_CHECKING: from torch import nn from transformers import PretrainedConfig, PreTrainedModel from ...hparams import ModelArguments if is_transformers_version_greater_than("4.57.0"): from transformers.models.qwen3_omni_moe import modeling_qwen3_omni_moe def _set_z3_leaf_modules(model: "PreTrainedModel", leaf_modules: list[Union["nn.Module", str]]) -> None: check_version("deepspeed>=0.13.0") from deepspeed.utils import set_z3_leaf_modules # type: ignore set_z3_leaf_modules(model, leaf_modules) def add_z3_leaf_module(model: "PreTrainedModel") -> None: r"""Set module as a leaf module to skip partitioning in deepspeed zero3.""" if not is_deepspeed_zero3_enabled(): return model_type = getattr(model.config, "model_type", None) text_config = getattr(model.config, "text_config", None) text_model_type = getattr(text_config, "model_type", None) if model_type == "dbrx": from transformers.models.dbrx.modeling_dbrx import DbrxFFN _set_z3_leaf_modules(model, [DbrxFFN]) if model_type == "deepseek_v2": # deepseek v2 uses custom code _set_z3_leaf_modules(model, ["DeepseekV2MoE"]) if model_type == "deepseek_v3" or model_type == "kimi_vl": # deepseek v3 and kimi vl use custom code _set_z3_leaf_modules(model, ["DeepseekV3MoE"]) if model_type == "ernie4_5_moe": from transformers.models.ernie4_5_moe.modeling_ernie4_5_moe import Ernie4_5_MoeSparseMoeBlock _set_z3_leaf_modules(model, [Ernie4_5_MoeSparseMoeBlock]) if model_type == "granitemoe": from transformers.models.granitemoe.modeling_granitemoe import GraniteMoeMoE _set_z3_leaf_modules(model, [GraniteMoeMoE]) if model_type == "glm4_moe": from transformers.models.glm4_moe.modeling_glm4_moe import Glm4MoeMoE _set_z3_leaf_modules(model, [Glm4MoeMoE]) if model_type == "glm4v_moe": from transformers.models.glm4v_moe.modeling_glm4v_moe import Glm4vMoeTextMoE _set_z3_leaf_modules(model, [Glm4vMoeTextMoE]) if model_type == "gpt_oss": from transformers.models.gpt_oss.modeling_gpt_oss import GptOssMLP _set_z3_leaf_modules(model, [GptOssMLP]) if model_type == "jamba": from transformers.models.jamba.modeling_jamba import JambaSparseMoeBlock _set_z3_leaf_modules(model, [JambaSparseMoeBlock]) if model_type == "jetmoe": from transformers.models.jetmoe.modeling_jetmoe import JetMoeMoA, JetMoeMoE _set_z3_leaf_modules(model, [JetMoeMoA, JetMoeMoE]) if model_type == "llama4": from transformers.models.llama4.modeling_llama4 import Llama4TextMoe _set_z3_leaf_modules(model, [Llama4TextMoe]) if model_type == "mixtral": from transformers.models.mixtral.modeling_mixtral import MixtralSparseMoeBlock _set_z3_leaf_modules(model, [MixtralSparseMoeBlock]) if model_type == "olmoe": from transformers.models.olmoe.modeling_olmoe import OlmoeSparseMoeBlock _set_z3_leaf_modules(model, [OlmoeSparseMoeBlock]) if model_type == "phimoe": from transformers.models.phimoe.modeling_phimoe import PhimoeSparseMoeBlock _set_z3_leaf_modules(model, [PhimoeSparseMoeBlock]) if model_type == "qwen2_moe": from transformers.models.qwen2_moe.modeling_qwen2_moe import Qwen2MoeSparseMoeBlock _set_z3_leaf_modules(model, [Qwen2MoeSparseMoeBlock]) if model_type == "qwen3_moe" or text_model_type == "qwen3_moe": # internvl 3.5 from transformers.models.qwen3_moe.modeling_qwen3_moe import Qwen3MoeSparseMoeBlock _set_z3_leaf_modules(model, [Qwen3MoeSparseMoeBlock]) if model_type == "qwen3_vl_moe": from transformers.models.qwen3_vl_moe.modeling_qwen3_vl_moe import Qwen3VLMoeTextSparseMoeBlock _set_z3_leaf_modules(model, [Qwen3VLMoeTextSparseMoeBlock]) if model_type in ("qwen3_omni_moe", "qwen3_omni_moe_thinker"): from transformers.models.qwen3_omni_moe.modeling_qwen3_omni_moe import Qwen3OmniMoeThinkerTextSparseMoeBlock _set_z3_leaf_modules(model, [Qwen3OmniMoeThinkerTextSparseMoeBlock]) def configure_moe(config: "PretrainedConfig", model_args: "ModelArguments", is_trainable: bool) -> None: if not is_trainable or not model_args.moe_aux_loss_coef: return model_type = getattr(config, "model_type", None) text_config = getattr(config, "text_config", None) # for multimodal model if model_type in [ "dbrx", "ernie4_5_moe", "granitemoe", "jamba", "jetmoe", "llama4", "mixtral", "olmoe", "phimoe", "qwen2_moe", "qwen3_moe", ]: setattr(config, "output_router_logits", True) if text_config and getattr(text_config, "model_type", None) in [ "glm4v_moe_text", # glmv4_5 "qwen3_moe", # internvl_3_5 ]: setattr(text_config, "output_router_logits", True) if model_type in [ "ernie4_5_moe", "granitemoe", "jamba", "llama4", "mixtral", "olmoe", "phimoe", "qwen2_moe", "qwen3_moe", ]: setattr(config, "router_aux_loss_coef", model_args.moe_aux_loss_coef) elif text_config and getattr(text_config, "model_type", None) in ["qwen3_moe"]: setattr(text_config, "router_aux_loss_coef", model_args.moe_aux_loss_coef) elif model_type == "deepseek": setattr(config, "aux_loss_alpha", model_args.moe_aux_loss_coef) elif model_type == "jetmoe": setattr(config, "aux_loss_coef", model_args.moe_aux_loss_coef) class Qwen3OmniMoeThinkerTextSparseMoeBlock(nn.Module): def __init__(self, config): super().__init__() self.num_experts = config.num_experts self.top_k = config.num_experts_per_tok self.norm_topk_prob = config.norm_topk_prob # gating self.gate = nn.Linear(config.hidden_size, config.num_experts, bias=False) self.experts = nn.ModuleList( [ modeling_qwen3_omni_moe.Qwen3OmniMoeThinkerTextMLP( config, intermediate_size=config.moe_intermediate_size ) for _ in range(self.num_experts) ] ) def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: batch_size, sequence_length, hidden_dim = hidden_states.shape hidden_states = hidden_states.view(-1, hidden_dim) # router_logits: (batch * sequence_length, n_experts) router_logits = self.gate(hidden_states) # Calculate the routing weights for all experts routing_weights = F.softmax(router_logits, dim=1, dtype=torch.float) # Retain the weight of the top_k and reset the rest of the expert rights to 0 (instead of retaining only top_k experts) top_k_weights, top_k_indices = torch.topk(routing_weights, self.top_k, dim=-1) # Initialize the all-zero weight matrix (same shape as all experts) full_routing_weights = torch.zeros_like(routing_weights) # Only the weight of top_k experts is retained, and the weight of the rest of the experts remains at 0 full_routing_weights.scatter_(1, top_k_indices, top_k_weights) # Normalized top_k weights (keep the original logic consistent) if self.norm_topk_prob: # Calculate the sum of the weights top_k each row (for normalization) top_k_sum = full_routing_weights.sum(dim=-1, keepdim=True) # Avoid dividing by zero top_k_sum = torch.clamp(top_k_sum, min=1e-9) full_routing_weights /= top_k_sum # Convert back to the input data type full_routing_weights = full_routing_weights.to(hidden_states.dtype) final_hidden_states = torch.zeros( (batch_size * sequence_length, hidden_dim), dtype=hidden_states.dtype, device=hidden_states.device ) # Go through all the experts (not just the selected ones) for expert_idx in range(self.num_experts): expert_layer = self.experts[expert_idx] # Get the weight of the current expert (inactive expert has a weight of 0 here) expert_weights = full_routing_weights[:, expert_idx, None] # shape: (batch*seq, 1) # All samples participate in the calculations of the current expert, the weight may be equal to 0 current_hidden_states = expert_layer(hidden_states) * expert_weights # Add-up to all expert outputs (experts with a weight of 0 do not affect the result) final_hidden_states += current_hidden_states final_hidden_states = final_hidden_states.reshape(batch_size, sequence_length, hidden_dim) return final_hidden_states, router_logits
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false
hiyouga/LlamaFactory
https://github.com/hiyouga/LlamaFactory/blob/f60a6e3d015962198b7c626936f117e83260bde9/src/llamafactory/model/model_utils/__init__.py
src/llamafactory/model/model_utils/__init__.py
python
Apache-2.0
f60a6e3d015962198b7c626936f117e83260bde9
2026-01-04T14:38:25.134237Z
false