Variations in battery usage scenarios and manufacturing differences can easily lead to variations in battery degradation data. This presents a challenge for data-driven methods, which must deal with inconsistent data distributions in the target and source domains. This is known as a cross-domain training prediction problem. Traditional data-driven methods may have limited generalization ability when dealing with such problems, leading to significant fluctuations in prediction results. Our solution is to map the source and target domains to a common
feature space and extract features from this space as input for prediction. This can help reduce variations in both domains.
电池使用场景和制造差异的变化可能导致电池退化数据的差异。这给数据驱动方法带来了挑战,这些方法必须处理目标域和源域中不一致的数据分布。这被称为跨域训练预测问题。在处理此类问题时,传统数据驱动方法可能具有有限的一般化能力,导致预测结果出现显著波动。我们的解决方案是将源域和目标域映射到公共特征空间,并从该空间提取特征作为预测的输入。这有助于减少两个域中的差异。2.2. Loss function 2.2. 损失函数
Transfer learning is a
machine learning technique where a model trained on one task is reused or adapted to another related task. This approach is particularly useful when there is limited data available for the target task, or when the target task is similar to the original task. Transfer learning usually includes the following modules:
迁移学习是一种机器学习技术,其中在一个任务上训练的模型被重用或调整以应用于另一个相关任务。这种方法在目标任务数据有限或目标任务与原始任务相似时特别有用。迁移学习通常包括以下模块:Feature extraction: The pre-trained model is used to extract relevant features from the input data, and these features are then used as input to a new model that is trained for the target task.
特征提取:使用预训练模型从输入数据中提取相关特征,然后将这些特征用作训练针对目标任务的新的模型的输入。
Domain adaptation: Domain adaptation is a principle that aims to enhance the performance of a machine learning model on a target domain by utilizing the knowledge from a source domain. The rationale for domain adaptation is that the data distribution of the target domain may vary from that of the source domain due to various factors, such as noise, bias, or temporal or spatial variations. This variation can impair the model's performance on the target domain. To tackle this problem, domain adaptation methods try to either minimize the difference between the source and target domains, or find a common representation that is robust to the domain change. There are different types of domain adaptation methods, depending on whether there are labeled or unlabeled data from the target domain. Supervised domain adaptation assumes that there are some labeled data from the target domain, but not enough to train a model from scratch. In this situation, the methods can either merge the source and target data to train a single model, or fine-tune a model trained on the source data using the target data. Unsupervised domain adaptation assumes that there are no labeled data from the target domain, only unlabeled data. In this situation, the methods can either use some measures to match the source and target data distributions, or learn a feature extractor that can map both domains to a shared space. Semi-supervised domain adaptation assumes that there are some labeled data and some unlabeled data from the target domain. In this situation, the methods can either use self-training or co-training techniques to label the unlabeled data and update the model, or use adversarial learning techniques to learn a domain
discriminator and a feature extractor that can trick it.
领域自适应:领域自适应是一种旨在通过利用源领域的知识来提高机器学习模型在目标领域性能的原则。领域自适应的原理是,由于各种因素,如噪声、偏差或时间或空间变化,目标域的数据分布可能与源域的数据分布不同。这种差异可能会损害模型在目标域的性能。为了解决这个问题,领域自适应方法试图最小化源域和目标域之间的差异,或者找到一个对领域变化具有鲁棒性的共同表示。根据目标域是否有标记或未标记的数据,存在不同类型的领域自适应方法。监督领域自适应假设目标域有一些标记数据,但不足以从头开始训练模型。在这种情况下,方法可以是合并源域和目标域的数据来训练单个模型,或者使用目标数据微调在源数据上训练的模型。 无监督领域自适应假设目标域没有标记数据,只有未标记数据。在这种情况下,方法可以使用一些措施来匹配源数据和目标数据分布,或者学习一个可以将两个领域映射到共享空间的特征提取器。半监督领域自适应假设目标域有一些标记数据和一些未标记数据。在这种情况下,方法可以使用自训练或协同训练技术来标记未标记数据并更新模型,或者使用对抗学习技术来学习一个领域判别器和可以欺骗它的特征提取器。Multi-task learning: The principle of multi-task learning involves training a single model to simultaneously perform multiple related tasks. The key idea is to exploit the shared information and relationships among tasks to improve overall performance. In multi-task learning, the model architecture typically consists of two main components: a shared feature representation and task-specific layers. The shared feature representation captures the common patterns and features across all tasks. It aims to learn a set of high-level representations that are beneficial for all tasks. The task-specific layers are responsible for mapping the shared representations to the specific outputs required by each task. During training, the model optimizes a
joint objective function that combines the losses from all the tasks. This objective function balances the contributions of each task and guides the model to learn a shared representation that benefits all tasks simultaneously. By jointly optimizing the model across tasks, the model can leverage the relationships and dependencies among tasks to improve performance. The benefits of multi-task learning arise from several factors. First, by learning from multiple tasks simultaneously, the model can access a larger and more diverse training dataset, which can improve generalization and robustness. Second, the shared representation allows the model to capture common knowledge and patterns, leading to more efficient learning. Third, multi-task learning enables the transfer of information between tasks, where knowledge gained from one task can aid in learning another task.
多任务学习:多任务学习的原理涉及训练一个单一模型同时执行多个相关任务。关键思想是利用任务之间的共享信息和关系来提高整体性能。在多任务学习中,模型架构通常由两个主要组件组成:共享特征表示和特定任务层。共享特征表示捕捉所有任务中的共同模式和特征。它旨在学习一组对所有任务都有益的高级表示。特定任务层负责将共享表示映射到每个任务所需的特定输出。在训练过程中,模型优化一个联合目标函数,该函数结合了所有任务的损失。此目标函数平衡每个任务的贡献,并指导模型学习一个对所有任务都有益的共享表示。通过跨任务联合优化模型,模型可以利用任务之间的关系和依赖来提高性能。多任务学习的优势源于几个因素。 首先,通过同时学习多个任务,模型可以访问更大、更多样化的训练数据集,从而提高泛化能力和鲁棒性。其次,共享表示允许模型捕捉共同知识和模式,导致更有效的学习。第三,多任务学习使得任务之间能够转移信息,其中一个任务中获得的知识可以帮助学习另一个任务。Each module generally requires an objective function for optimization. The optimization function proposed in this paper is as follows:
每个模块通常需要一个目标函数进行优化。本文提出的优化函数如下:- (1)
Domain adaptation 领域自适应
Due to differences in battery usage scenarios, the distribution of features in the source and target domains can often vary. To measure the difference between these distributions, we use multiple kernel maximum mean discrepancy (MK-MMD).
由于电池使用场景的不同,源域和目标域中功能的分布往往会有所差异。为了衡量这些分布之间的差异,我们使用多核最大均值差异(MK-MMD)。
Let
and
represent the marginal distributions of the features in the source and target domains, respectively. The maximum mean discrepancy (MMD) is defined as:
设 和 分别代表源域和目标域中特征的边缘分布。最大均值差异(MMD)定义为:(1)where
and
are the expected operation of the edge distribution of
and
, respectively.
denotes represents the 2-norm operation of the reproducing kernel Hilbert space (RKHS). is the mapping function for the source domain and target domain feature matrices to RKHS.
和 分别表示 和 的边缘分布的预期操作。 表示再生核希尔伯特空间(RKHS)的 2-范数操作。 是将源域和目标域特征矩阵映射到 RKHS 的映射函数。Due to the complexity of calculating the mean value of all features in the source and target domains, we approximate Eq.
(1) using the mean values of samples in the source and target domains.
由于计算源域和目标域所有特征均值复杂,我们使用源域和目标域样本的均值来近似公式(1)。(2)The inner product of mapping function
is expressed by kernel function, and Eq.
(2) can be expressed as.
映射函数 的内积由核函数表示,式(2)可以表示为。(3)where
哪里 is the kernel matrix computed by the Gaussian kernel. is the Gaussian kernel mapping of each sample.
是通过高斯核计算得到的核矩阵。 是每个样本的高斯核映射。- (2)
Prediction losses 预测损失
In this paper, we use the root mean square error (MSE) as the loss function for prediction results. This includes the loss for battery discharge capacity (as shown in Eq. (4)) and the loss for remaining useful life (as shown in Eq.
(5)).
在这篇论文中,我们使用均方根误差(MSE)作为预测结果的损失函数。这包括电池放电容量损失(如图(4)所示)和剩余使用寿命损失(如图(5)所示)。(4)(5) 2.3. Multi-task objective function weight adaptive optimization
2.3. 多任务目标函数权重自适应优化
In multi-task learning, model training typically involves weighting multiple loss functions to obtain an overall loss. However, accurately quantifying the magnitude of the loss function used by different tasks and the importance of each task can be difficult. Many existing studies choose an optimal value through trial and error, which can be time-consuming. Others use industry-wide parameters, but whether these are optimal remains to be proven. One potential solution is to automatically adjust the weight of each loss function to improve training efficiency and potentially optimize model performance.
在多任务学习中,模型训练通常涉及对多个损失函数进行加权以获得总体损失。然而,准确量化不同任务所使用的损失函数的幅度以及每个任务的重要性可能很困难。许多现有研究通过试错选择最优值,这可能很耗时。其他人使用行业参数,但这些是否最优尚待证明。一个潜在的解决方案是自动调整每个损失函数的权重,以提高训练效率并可能优化模型性能。
Assuming that the multi-task loss function is represented by and the corresponding weight is represented by , the overall loss function can be expressed as:
假设多任务损失函数由 表示,相应的权重由 表示,整体损失函数可以表示为:(6)
Battery health status prediction belongs to regression task. Any task can be defined as
电池健康状态预测属于回归任务。任何任务都可以定义为(7)where is the network prediction value. The above equation represents a multivariate Gaussian distribution with as the mean vector and as the variance.
是网络预测值。上述方程表示一个以 为均值向量、 为方差的多元高斯分布。
The logarithmic likelihood of the above equation is
对上述方程的对数似然为(8)
Define the regression loss as . According to the above equation, the loss function of the overall training task can be expressed as.
定义回归损失为 。根据上述方程,整体训练任务的损失函数可以表示为。(9)
The goal of the training is to minimize this maximum likelihood estimate. As increases, the corresponding weight will decrease. Conversely, as
decreases, the corresponding weight will increase.
训练的目标是使最大似然估计最小化。当 增加时,相应的权重将减少。相反,当 减少时,相应的权重将增加。