本研究提出 HILA-Merging,一種無需訓練即可在共享的大型語言模型(LLM)架 構中整合多語言與多領域專家知識的方法。此方法旨在解決跨語言與跨領域專業能力 的遷移難題,同時避免多階段微調所帶來的高昂計算成本。HILA 透過細粒度、基於 loss 的結構分析運作。本研究的核心是以層級與頭級為單位的重要度量測機制,利用專 家模型生成的黃金資料集,量化每個注意力頭與 MLP 層的功能性貢獻。這些重要度訊 號同時引導剪枝與自適應加權融合,使方法能選擇性地保留關鍵專家組件,同時降低 不同專家模型之間的干擾。 我們在醫療、程式設計與金融三個領域、以及多語言設定下評估本方法。結果顯 示,HILA-Merging 持續優於基於幅度或啟發式的模型融合方法。在醫療領域的基準測 試中,我們的方法在 MedQA 上超越最強基線最多 1.65 分,並在 MedMCQA 上提升 1.6 分。在程式設計任務中,我們的方法在 HumanEval-XL 上提升超過 6 分,展現對結 構化推理任務的明顯增益。在金融領域中,HILA 在 II-Finance 上帶來約 0.4 至 0.6 分 的提升,同時在 FinanceQA 上維持具競爭力的表現並保留專家能力。這些結果整體顯 示,基於重要度、無需訓練的融合策略能在異質任務中帶來穩健的效能提升,並有效 支持跨語言知識遷移。 綜合上述發現,基於損失的結構引導為構建多語言、多領域的大型語言模型提供 了一條具可擴展性與良好泛化能力的途徑,而無需額外訓練。本研究之原始碼已公開 於:https://github.com/charlie963258/HILA-Merging;This work introduces HILA-Merging, a training-free method for integrating multilin- gual and domain-specialist knowledge within a shared LLM architecture. Our method addresses the challenge of transferring expertise across languages and domains without incurring the computational cost of multi-stage fine-tuning. HILA operates through fine- grained, loss-aware structural analysis. At the core of our work is a layerwise and head- wise importance measurement procedure that quantifies the functional contribution of each attention head and MLP layer using specialist-generated golden datasets. These im- portance signals guide both pruning and adaptive weighted merging, enabling our method to selectively preserve critical expert components while mitigating interference between specialists. We evaluate our method across medical, programming, and finance domains under multilingual settings. It demonstrates consistent improvements over magnitude-based and heuristic merging methods. In medical benchmarks, our method exceeds the strongest baseline by up to 1.65 points on MedQA and 1.6 points on MedMCQA. In programming, our method improves HumanEval-XL accuracy by more than 6 points, showing substantial gains in structured reasoning tasks. In finance, HILA improves II-Finance performance by approximately 0.4 to 0.6 points, while maintaining competitive results on FinanceQA and preserving specialist knowledge. These results collectively show that importance-guided, training-free merging provides reliable performance gains across heterogeneous tasks and supports effective cross-lingual transfer. Together, the findings indicate that loss-aware structural guidance offers a scalable and generalizable path toward building multilingual, multi-domain LLMs without additional training. We release the source code at: https://github.com/charlie963258/HILA- Merging