improve finetune guide
This commit is contained in:
parent
55210c89e2
commit
7210a7481e
@ -223,5 +223,11 @@
|
|||||||
"Merge model successfully": "合并模型成功",
|
"Merge model successfully": "合并模型成功",
|
||||||
"Convert Data successfully": "数据转换成功",
|
"Convert Data successfully": "数据转换成功",
|
||||||
"Please select a LoRA model": "请选择一个LoRA模型",
|
"Please select a LoRA model": "请选择一个LoRA模型",
|
||||||
"You are using sample data for training. For formal training, please make sure to create your own jsonl file.": "你正在使用示例数据训练,对于正式训练场合,请务必创建你自己的jsonl训练数据"
|
"You are using sample data for training. For formal training, please make sure to create your own jsonl file.": "你正在使用示例数据训练,对于正式训练场合,请务必创建你自己的jsonl训练数据",
|
||||||
|
"WSL is not running. You may be using an outdated version of WSL, run \"wsl --update\" to update.": "WSL没有运行。你可能正在使用旧版本的WSL,请在cmd执行\"wsl --update\"以更新",
|
||||||
|
"Memory is not enough, try to increase the virtual memory or use a smaller base model.": "内存不足,尝试增加虚拟内存,或使用一个更小规模的基底模型",
|
||||||
|
"VRAM is not enough": "显存不足",
|
||||||
|
"Training data is not enough, reduce context length or add more data for training": "训练数据不足,请减小上下文长度或增加训练数据",
|
||||||
|
"You are using WSL 1 for training, please upgrade to WSL 2. e.g. Run \"wsl --set-version Ubuntu-22.04 2\"": "你正在使用WSL 1进行训练,请升级到WSL 2。例如,运行\"wsl --set-version Ubuntu-22.04 2\"",
|
||||||
|
"Matched CUDA is not installed": "未安装匹配的CUDA"
|
||||||
}
|
}
|
@ -37,6 +37,7 @@ import {
|
|||||||
import { Line } from 'react-chartjs-2';
|
import { Line } from 'react-chartjs-2';
|
||||||
import { ChartJSOrUndefined } from 'react-chartjs-2/dist/types';
|
import { ChartJSOrUndefined } from 'react-chartjs-2/dist/types';
|
||||||
import { WindowShow } from '../../wailsjs/runtime';
|
import { WindowShow } from '../../wailsjs/runtime';
|
||||||
|
import { t } from 'i18next';
|
||||||
|
|
||||||
ChartJS.register(
|
ChartJS.register(
|
||||||
CategoryScale,
|
CategoryScale,
|
||||||
@ -52,12 +53,13 @@ const parseLossData = (data: string) => {
|
|||||||
const regex = /Epoch (\d+):\s+(\d+%)\|[\s\S]*\| (\d+)\/(\d+) \[(\d+:\d+)<(\d+:\d+),\s+(\d+.\d+it\/s), loss=(\S+),[\s\S]*\]/g;
|
const regex = /Epoch (\d+):\s+(\d+%)\|[\s\S]*\| (\d+)\/(\d+) \[(\d+:\d+)<(\d+:\d+),\s+(\d+.\d+it\/s), loss=(\S+),[\s\S]*\]/g;
|
||||||
const matches = Array.from(data.matchAll(regex));
|
const matches = Array.from(data.matchAll(regex));
|
||||||
if (matches.length === 0)
|
if (matches.length === 0)
|
||||||
return;
|
return false;
|
||||||
const lastMatch = matches[matches.length - 1];
|
const lastMatch = matches[matches.length - 1];
|
||||||
const epoch = parseInt(lastMatch[1]);
|
const epoch = parseInt(lastMatch[1]);
|
||||||
const loss = parseFloat(lastMatch[8]);
|
const loss = parseFloat(lastMatch[8]);
|
||||||
commonStore.setChartTitle(`Epoch ${epoch}: ${lastMatch[2]} - ${lastMatch[3]}/${lastMatch[4]} - ${lastMatch[5]}/${lastMatch[6]} - ${lastMatch[7]} Loss=${loss}`);
|
commonStore.setChartTitle(`Epoch ${epoch}: ${lastMatch[2]} - ${lastMatch[3]}/${lastMatch[4]} - ${lastMatch[5]}/${lastMatch[6]} - ${lastMatch[7]} Loss=${loss}`);
|
||||||
addLossDataToChart(epoch, loss);
|
addLossDataToChart(epoch, loss);
|
||||||
|
return true;
|
||||||
};
|
};
|
||||||
|
|
||||||
let chartLine: ChartJSOrUndefined<'line', (number | null)[], string>;
|
let chartLine: ChartJSOrUndefined<'line', (number | null)[], string>;
|
||||||
@ -140,10 +142,36 @@ const loraFinetuneParametersOptions: Array<[key: keyof LoraFinetuneParameters, t
|
|||||||
['headQk', 'boolean', 'Head QK']
|
['headQk', 'boolean', 'Head QK']
|
||||||
];
|
];
|
||||||
|
|
||||||
|
const showError = (e: any) => {
|
||||||
|
const msg = e.message || e;
|
||||||
|
if (msg === 'wsl not running') {
|
||||||
|
toast(t('WSL is not running. You may be using an outdated version of WSL, run "wsl --update" to update.'), { type: 'error' });
|
||||||
|
} else {
|
||||||
|
toast(t(msg), { type: 'error' });
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
const errorsMap = Object.entries({
|
||||||
|
'killed python3 ./finetune/lora/train.py': 'Memory is not enough, try to increase the virtual memory or use a smaller base model.',
|
||||||
|
'cuda out of memory': 'VRAM is not enough',
|
||||||
|
'valueerror: high <= 0': 'Training data is not enough, reduce context length or add more data for training',
|
||||||
|
'+= \'+ptx\'': 'You are using WSL 1 for training, please upgrade to WSL 2. e.g. Run "wsl --set-version Ubuntu-22.04 2"',
|
||||||
|
'cuda_home environment variable is not set': 'Matched CUDA is not installed',
|
||||||
|
'unsupported gpu architecture': 'Matched CUDA is not installed',
|
||||||
|
'error building extension \'fused_adam\'': 'Matched CUDA is not installed'
|
||||||
|
});
|
||||||
|
|
||||||
export const wslHandler = (data: string) => {
|
export const wslHandler = (data: string) => {
|
||||||
if (data) {
|
if (data) {
|
||||||
addWslMessage(data);
|
addWslMessage(data);
|
||||||
parseLossData(data);
|
const ok = parseLossData(data);
|
||||||
|
if (!ok)
|
||||||
|
for (const [key, value] of errorsMap) {
|
||||||
|
if (data.toLowerCase().includes(key)) {
|
||||||
|
showError(value);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
@ -188,12 +216,8 @@ const Terminal: FC = observer(() => {
|
|||||||
WslStart().then(() => {
|
WslStart().then(() => {
|
||||||
addWslMessage('WSL> ' + input);
|
addWslMessage('WSL> ' + input);
|
||||||
setInput('');
|
setInput('');
|
||||||
WslCommand(input).catch((e: any) => {
|
WslCommand(input).catch(showError);
|
||||||
toast((e.message || e), { type: 'error' });
|
}).catch(showError);
|
||||||
});
|
|
||||||
}).catch((e: any) => {
|
|
||||||
toast((e.message || e), { type: 'error' });
|
|
||||||
});
|
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
@ -208,9 +232,7 @@ const Terminal: FC = observer(() => {
|
|||||||
<Button onClick={() => {
|
<Button onClick={() => {
|
||||||
WslStop().then(() => {
|
WslStop().then(() => {
|
||||||
toast(t('Command Stopped'), { type: 'success' });
|
toast(t('Command Stopped'), { type: 'success' });
|
||||||
}).catch((e: any) => {
|
}).catch(showError);
|
||||||
toast((e.message || e), { type: 'error' });
|
|
||||||
});
|
|
||||||
}}>
|
}}>
|
||||||
{t('Stop')}
|
{t('Stop')}
|
||||||
</Button>
|
</Button>
|
||||||
@ -256,7 +278,7 @@ const LoraFinetune: FC = observer(() => {
|
|||||||
if (!ok)
|
if (!ok)
|
||||||
return;
|
return;
|
||||||
|
|
||||||
const convertedDataPath = `./finetune/json2binidx_tool/data/${dataParams.dataPath.split('/').pop()!.split('.')[0]}_text_document`;
|
const convertedDataPath = `./finetune/json2binidx_tool/data/${dataParams.dataPath.split(/[\/\\]/).pop()!.split('.')[0]}_text_document`;
|
||||||
if (!await FileExists(convertedDataPath + '.idx')) {
|
if (!await FileExists(convertedDataPath + '.idx')) {
|
||||||
toast(t('Please convert data first.'), { type: 'error' });
|
toast(t('Please convert data first.'), { type: 'error' });
|
||||||
return;
|
return;
|
||||||
@ -302,9 +324,7 @@ const LoraFinetune: FC = observer(() => {
|
|||||||
`--beta1 ${loraParams.beta1} --beta2 ${loraParams.beta2} --adam_eps ${loraParams.adamEps} ` +
|
`--beta1 ${loraParams.beta1} --beta2 ${loraParams.beta2} --adam_eps ${loraParams.adamEps} ` +
|
||||||
`--devices ${loraParams.devices} --precision ${loraParams.precision} ` +
|
`--devices ${loraParams.devices} --precision ${loraParams.precision} ` +
|
||||||
`--grad_cp ${loraParams.gradCp ? '1' : '0'} ` +
|
`--grad_cp ${loraParams.gradCp ? '1' : '0'} ` +
|
||||||
`--lora_r ${loraParams.loraR} --lora_alpha ${loraParams.loraAlpha} --lora_dropout ${loraParams.loraDropout}`).catch((e: any) => {
|
`--lora_r ${loraParams.loraR} --lora_alpha ${loraParams.loraAlpha} --lora_dropout ${loraParams.loraDropout}`).catch(showError);
|
||||||
toast((e.message || e), { type: 'error' });
|
|
||||||
});
|
|
||||||
}).catch(e => {
|
}).catch(e => {
|
||||||
const msg = e.message || e;
|
const msg = e.message || e;
|
||||||
if (msg === 'ubuntu not found') {
|
if (msg === 'ubuntu not found') {
|
||||||
@ -332,9 +352,7 @@ const LoraFinetune: FC = observer(() => {
|
|||||||
type: 'info',
|
type: 'info',
|
||||||
autoClose: false
|
autoClose: false
|
||||||
});
|
});
|
||||||
}).catch(e => {
|
}).catch(showError);
|
||||||
toast((e.message || e), { type: 'error' });
|
|
||||||
});
|
|
||||||
});
|
});
|
||||||
};
|
};
|
||||||
|
|
||||||
@ -343,7 +361,7 @@ const LoraFinetune: FC = observer(() => {
|
|||||||
} else if (msg.includes('wsl.state: The system cannot find the file')) {
|
} else if (msg.includes('wsl.state: The system cannot find the file')) {
|
||||||
enableWsl(true);
|
enableWsl(true);
|
||||||
} else {
|
} else {
|
||||||
toast(msg, { type: 'error' });
|
showError(msg);
|
||||||
}
|
}
|
||||||
});
|
});
|
||||||
};
|
};
|
||||||
@ -399,14 +417,15 @@ const LoraFinetune: FC = observer(() => {
|
|||||||
onChange={(e, data) => {
|
onChange={(e, data) => {
|
||||||
setDataParams({ vocabPath: data.value });
|
setDataParams({ vocabPath: data.value });
|
||||||
}} />
|
}} />
|
||||||
<Button appearance="secondary" size="large" onClick={() => {
|
<Button appearance="secondary" size="large" onClick={async () => {
|
||||||
|
const ok = await checkDependencies(navigate);
|
||||||
|
if (!ok)
|
||||||
|
return;
|
||||||
ConvertData(commonStore.settings.customPythonPath, dataParams.dataPath,
|
ConvertData(commonStore.settings.customPythonPath, dataParams.dataPath,
|
||||||
'./finetune/json2binidx_tool/data/' + dataParams.dataPath.split('/').pop()!.split('.')[0],
|
'./finetune/json2binidx_tool/data/' + dataParams.dataPath.split(/[\/\\]/).pop()!.split('.')[0],
|
||||||
dataParams.vocabPath).then(() => {
|
dataParams.vocabPath).then(() => {
|
||||||
toast(t('Convert Data successfully'), { type: 'success' });
|
toast(t('Convert Data successfully'), { type: 'success' });
|
||||||
}).catch((e: any) => {
|
}).catch(showError);
|
||||||
toast((e.message || e), { type: 'error' });
|
|
||||||
});
|
|
||||||
}}>{t('Convert')}</Button>
|
}}>{t('Convert')}</Button>
|
||||||
</div>
|
</div>
|
||||||
</div>
|
</div>
|
||||||
@ -458,9 +477,7 @@ const LoraFinetune: FC = observer(() => {
|
|||||||
`models/${loraParams.baseModel}-LoRA-${loraParams.loraLoad}`).then(() => {
|
`models/${loraParams.baseModel}-LoRA-${loraParams.loraLoad}`).then(() => {
|
||||||
toast(t('Merge model successfully'), { type: 'success' });
|
toast(t('Merge model successfully'), { type: 'success' });
|
||||||
refreshLocalModels({ models: commonStore.modelSourceList }, false);
|
refreshLocalModels({ models: commonStore.modelSourceList }, false);
|
||||||
}).catch((e: any) => {
|
}).catch(showError);
|
||||||
toast((e.message || e), { type: 'error' });
|
|
||||||
});
|
|
||||||
} else {
|
} else {
|
||||||
toast(t('Please select a LoRA model'), { type: 'info' });
|
toast(t('Please select a LoRA model'), { type: 'info' });
|
||||||
}
|
}
|
||||||
@ -522,9 +539,7 @@ const LoraFinetune: FC = observer(() => {
|
|||||||
<Button appearance="secondary" size="large" onClick={() => {
|
<Button appearance="secondary" size="large" onClick={() => {
|
||||||
WslStop().then(() => {
|
WslStop().then(() => {
|
||||||
toast(t('Command Stopped'), { type: 'success' });
|
toast(t('Command Stopped'), { type: 'success' });
|
||||||
}).catch((e: any) => {
|
}).catch(showError);
|
||||||
toast((e.message || e), { type: 'error' });
|
|
||||||
});
|
|
||||||
}}>{t('Stop')}</Button>
|
}}>{t('Stop')}</Button>
|
||||||
<Button appearance="primary" size="large" onClick={StartLoraFinetune}>{t('Train')}</Button>
|
<Button appearance="primary" size="large" onClick={StartLoraFinetune}>{t('Train')}</Button>
|
||||||
</div>
|
</div>
|
||||||
|
Loading…
x
Reference in New Issue
Block a user