[go: up one dir, main page]

CN111291868A - Network model training method, device, equipment and computer readable storage medium - Google Patents

Network model training method, device, equipment and computer readable storage medium Download PDF

Info

Publication number
CN111291868A
CN111291868A CN202010145962.1A CN202010145962A CN111291868A CN 111291868 A CN111291868 A CN 111291868A CN 202010145962 A CN202010145962 A CN 202010145962A CN 111291868 A CN111291868 A CN 111291868A
Authority
CN
China
Prior art keywords
network
training
sub
data
candidate sub
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010145962.1A
Other languages
Chinese (zh)
Other versions
CN111291868B (en
Inventor
王达
刘博�
郑文琛
杨强
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
WeBank Co Ltd
Original Assignee
WeBank Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by WeBank Co Ltd filed Critical WeBank Co Ltd
Priority to CN202010145962.1A priority Critical patent/CN111291868B/en
Publication of CN111291868A publication Critical patent/CN111291868A/en
Application granted granted Critical
Publication of CN111291868B publication Critical patent/CN111291868B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Software Systems (AREA)
  • Computing Systems (AREA)
  • Artificial Intelligence (AREA)
  • Mathematical Physics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Medical Informatics (AREA)
  • Data Exchanges In Wide-Area Networks (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

本发明公开了一种网络模型训练方法,包括以下步骤:获取预设消息对应的训练数据以及验证数据;通过深度神经子网络生成器生成多个候选子网络,并基于所述训练数据以及验证数据,确定各个候选子网络对应的训练误差;基于各个所述训练误差,在各个候选子网络中确定最优子网络,并基于所述最优子网络以及主网络确定目标预测模型。本发明还公开了一种网络模型训练装置、设备及计算机可读存储介质。本发明能够通过机器学习实现网络模型的自动训练,针对不同的预设消息,无需人工进行不同网络模型的训练和优化,节省了大量的人力资源,降低了模型训练与调优的工作量,提高了网络模型的训练效率。

Figure 202010145962

The invention discloses a network model training method, comprising the following steps: acquiring training data and verification data corresponding to preset messages; generating multiple candidate sub-networks through a deep neural sub-network generator, and based on the training data and verification data , determine the training error corresponding to each candidate sub-network; based on each of the training errors, determine the optimal sub-network in each candidate sub-network, and determine the target prediction model based on the optimal sub-network and the main network. The invention also discloses a network model training device, equipment and computer-readable storage medium. The present invention can realize the automatic training of network models through machine learning. For different preset messages, there is no need to manually train and optimize different network models, save a lot of human resources, reduce the workload of model training and optimization, and improve the The training efficiency of the network model.

Figure 202010145962

Description

网络模型训练方法、装置、设备及计算机可读存储介质Network model training method, apparatus, device, and computer-readable storage medium

技术领域technical field

本发明涉及人工智能技术领域,尤其涉及一种网络模型训练方法、装置、设备及计算机可读存储介质。The present invention relates to the technical field of artificial intelligence, and in particular, to a network model training method, apparatus, device, and computer-readable storage medium.

背景技术Background technique

随着人工智能技术的快速发展,神经网络模型(可以简称为网络模型)在系统辨识、模式识别、智能控制等领域有着广泛的应用前景。通常可以基于训练样本集,对预设的网络模型进行训练,得到训练好的网络模型,进而,可以将待检测样本输入至训练好的网络模型,得到网络模型的实际输出结果,实际输出结果为对待检测样本进行预测的预测结果。With the rapid development of artificial intelligence technology, neural network models (which can be referred to as network models) have broad application prospects in the fields of system identification, pattern recognition, and intelligent control. Usually, a preset network model can be trained based on the training sample set to obtain a trained network model, and then the samples to be tested can be input into the trained network model to obtain the actual output result of the network model. The actual output result is The prediction result of the prediction of the sample to be tested.

例如,在程序化广告投放中,可以将每个用户的特点输入至网络模型根据,根据网络模型的实际输出结果快速做出是否投放广告的决策。For example, in programmatic advertising, the characteristics of each user can be input into the network model, and a decision on whether to place an advertisement can be made quickly according to the actual output of the network model.

但是,目前采用的网络模型需要专业的算法工程师结合不同广告业务的特点有针对性地进行模型训练和优化,不同的广告业务无法使用相同的网络模型,算法工程师需要针对不同的广告业务进行不同网络模型的训练和优化,需要花费算法工程师的大量时间,并且每次训练所采用的模型结构需要算法人员手动指定,导致不同网络模型的训练和优化不够智能。However, the currently adopted network model requires professional algorithm engineers to conduct model training and optimization based on the characteristics of different advertising services. Different advertising services cannot use the same network model, and algorithm engineers need to conduct different network models for different advertising services. The training and optimization of the model requires a lot of time for algorithm engineers, and the model structure used in each training needs to be manually specified by the algorithm personnel, resulting in the training and optimization of different network models are not intelligent enough.

上述内容仅用于辅助理解本发明的技术方案,并不代表承认上述内容是现有技术。The above content is only used to assist the understanding of the technical solutions of the present invention, and does not mean that the above content is the prior art.

发明内容SUMMARY OF THE INVENTION

本发明的主要目的在于提供一种网络模型训练方法、装置、设备及计算机可读存储介质,旨在解决不同网络模型的训练和优化不智能的技术问题。The main purpose of the present invention is to provide a network model training method, device, equipment and computer-readable storage medium, aiming at solving the technical problem of unintelligent training and optimization of different network models.

为实现上述目的,本发明提供一种网络模型训练方法,所述网络模型训练方法包括以下步骤:In order to achieve the above object, the present invention provides a network model training method, the network model training method includes the following steps:

获取消息对应的训练数据以及验证数据;Obtain the training data and verification data corresponding to the message;

通过深度神经子网络生成器生成多个候选子网络,并基于所述训练数据以及验证数据,确定各个候选子网络对应的训练误差;Generate a plurality of candidate sub-networks by the deep neural sub-network generator, and determine the training error corresponding to each candidate sub-network based on the training data and the verification data;

基于各个所述训练误差,在各个候选子网络中确定最优子网络,并基于所述最优子网络以及主网络确定目标预测模型。Based on each of the training errors, an optimal sub-network is determined in each candidate sub-network, and a target prediction model is determined based on the optimal sub-network and the main network.

进一步地,所述基于所述训练数据以及验证数据,确定各个候选子网络对应的训练误差的步骤包括:Further, the step of determining the training error corresponding to each candidate sub-network based on the training data and the verification data includes:

依次在各个候选子网络获取目标候选子网络;Obtain target candidate sub-networks in each candidate sub-network in turn;

将所述目标候选子网络嵌入主网络得到更新后的主网络,并将所述训练数据以及验证数据输入更新后的主网络,以获得所述目标候选子网络对应的训练误差。Embed the target candidate sub-network into the main network to obtain an updated main network, and input the training data and verification data into the updated main network to obtain the training error corresponding to the target candidate sub-network.

进一步地,所述将所述训练数据以及验证数据输入更新后的主网络,以获得所述目标候选子网络对应的训练误差的步骤包括:Further, the step of inputting the training data and verification data into the updated main network to obtain the training error corresponding to the target candidate sub-network includes:

将所述训练数据以及验证数据输入更新后的主网络,并更新训练次数;Input the training data and the verification data into the updated main network, and update the training times;

在所述训练次数达到预设训练次数时,将训练后的更新后的主网络,对应的训练误差作为所述目标候选子网络对应的训练误差;When the number of training times reaches the preset number of training times, the training error corresponding to the updated main network after training is used as the training error corresponding to the target candidate sub-network;

在所述训练次数未达到预设训练次数时,将训练后的更新后的主网络作为更新后的主网络,并返回将所述训练数据以及验证数据输入更新后的主网络,并更新训练次数的步骤。When the number of trainings does not reach the preset number of trainings, the updated main network after training is used as the updated main network, and the training data and verification data are input into the updated main network, and the training times are updated. A step of.

进一步地,所述基于各个所述训练误差,在各个候选子网络中确定最优子网络的步骤包括:Further, the step of determining the optimal sub-network in each candidate sub-network based on each of the training errors includes:

获取各个候选子网络的网络层数对应的惩罚系数;Obtain the penalty coefficient corresponding to the number of network layers of each candidate sub-network;

基于各个惩罚系数以及训练误差,分别确定各个候选子网络对应的权重分值;Based on each penalty coefficient and training error, the corresponding weight scores of each candidate sub-network are determined respectively;

将各个候选子网络中权重分值最小的候选子网络作为所述最优子网络。The candidate sub-network with the smallest weight score among the candidate sub-networks is used as the optimal sub-network.

进一步地,所述网络模型训练方法还包括:Further, the network model training method also includes:

在通过深度神经子网络生成器生成多个候选子网络时,更新搜索次数;When multiple candidate sub-networks are generated by the deep neural sub-network generator, update the number of searches;

所述基于所述最优子网络以及主网络确定所述目标预测模型的步骤包括:The step of determining the target prediction model based on the optimal sub-network and the main network includes:

确定所述最优子网络对应的训练误差是否小于所述主网络对应的训练误差;Determine whether the training error corresponding to the optimal sub-network is smaller than the training error corresponding to the main network;

若所述最优子网络对应的训练误差小于所述主网络对应的训练误差,则确定所述搜索次数是否达到预设搜索次数;If the training error corresponding to the optimal sub-network is smaller than the training error corresponding to the main network, determine whether the number of searches reaches a preset number of searches;

在所述搜索次数达到所述预设搜索次数时,则将所述最优子网络嵌入所述主网络以获得所述目标预测模型。When the number of searches reaches the preset number of searches, the optimal sub-network is embedded in the main network to obtain the target prediction model.

进一步地,所述确定所述最优子网络对应的训练误差是否小于所述主网络对应的训练误差的步骤之后,还包括:Further, after the step of determining whether the training error corresponding to the optimal sub-network is smaller than the training error corresponding to the main network, the method further includes:

若所述最优子网络对应的训练误差大于或等于所述主网络对应的训练误差,则返回执行通过深度神经子网络生成器生成多个候选子网络的步骤。If the training error corresponding to the optimal sub-network is greater than or equal to the training error corresponding to the main network, return to the step of generating multiple candidate sub-networks through the deep neural sub-network generator.

进一步地,所述确定所述搜索次数是否达到预设搜索次数的步骤之后,还包括:Further, after the step of determining whether the number of searches reaches a preset number of searches, the method further includes:

在所述搜索次数未达到所述预设搜索次数时,则将所述最优子网络嵌入所述主网络;When the number of searches does not reach the preset number of searches, the optimal sub-network is embedded in the main network;

将嵌入所述最优子网络后的主网络作为所述主网络,并返回执行通过深度神经子网络生成器生成多个候选子网络的步骤。The main network after embedding the optimal sub-network is used as the main network, and the step of generating a plurality of candidate sub-networks by the deep neural sub-network generator is performed back.

进一步地,所述基于所述最优子网络以及主网络确定目标预测模型的步骤之后,所述网络模型训练方法还包括:Further, after the step of determining the target prediction model based on the optimal sub-network and the main network, the network model training method further includes:

在确定目标预测模型之后的持续时长达到预设持续时长时,返回执行获取预设消息对应的训练数据以及验证数据的步骤。When the duration after the target prediction model is determined reaches the preset duration, the steps of obtaining training data and verification data corresponding to the preset message are returned to.

进一步地,所述训练数据为用户特征数据,所述验证数据为曝光点击数据,所述基于所述最优子网络以及主网络确定目标预测模型的步骤之后,所述网络模型训练还包括:Further, the training data is user feature data, the verification data is exposure click data, and after the step of determining the target prediction model based on the optimal sub-network and the main network, the network model training further includes:

获取待预测用户特征数据,并将所述待预测用户数据输入所述目标预测模型,以获得预设消息对应的点击概率;Obtaining the feature data of the user to be predicted, and inputting the user data to be predicted into the target prediction model to obtain the click probability corresponding to the preset message;

在所述点击概率大于预设阈值时,确定所述待预测用户特征数据对应的用户为待投放用户。When the click probability is greater than a preset threshold, it is determined that the user corresponding to the to-be-predicted user characteristic data is the to-be-delivered user.

此外,为实现上述目的,本发明还提供一种网络模型训练装置,所述网络模型训练装置包括:In addition, in order to achieve the above object, the present invention also provides a network model training device, the network model training device includes:

获取模块,用于获取预设消息对应的训练数据以及验证数据;an acquisition module for acquiring training data and verification data corresponding to the preset message;

生成模块,用于通过深度神经子网络生成器生成多个候选子网络,并基于所述训练数据以及验证数据,确定各个候选子网络对应的训练误差;A generation module, used for generating multiple candidate sub-networks through the deep neural sub-network generator, and determining the training error corresponding to each candidate sub-network based on the training data and the verification data;

确定模块,用于基于各个所述训练误差,在各个候选子网络中确定最优子网络,接着基于所述最优子网络以及主网络确定目标预测模型。A determination module, configured to determine an optimal sub-network in each candidate sub-network based on each of the training errors, and then determine a target prediction model based on the optimal sub-network and the main network.

此外,为实现上述目的,本发明还提供一种网络模型训练设备,所述网络模型训练设备包括:存储器、处理器及存储在所述存储器上并可在所述处理器上运行的网络模型训练程序,所述网络模型训练程序被所述处理器执行时实现前述的网络模型训练方法的步骤。In addition, in order to achieve the above object, the present invention also provides a network model training device, the network model training device includes: a memory, a processor, and a network model training device stored on the memory and running on the processor A program, when the network model training program is executed by the processor, implements the steps of the aforementioned network model training method.

此外,为实现上述目的,本发明还提供一种计算机可读存储介质,所述计算机可读存储介质上存储有网络模型训练程序,所述网络模型训练程序被处理器执行时实现前述的网络模型训练方法的步骤。In addition, in order to achieve the above object, the present invention also provides a computer-readable storage medium, where a network model training program is stored on the computer-readable storage medium, and when the network model training program is executed by a processor, the aforementioned network model is implemented The steps of the training method.

本发明通过获取预设消息对应的训练数据以及验证数据;接着通过深度神经子网络生成器生成多个候选子网络,并基于所述训练数据以及验证数据,确定各个候选子网络对应的训练误差;而后基于各个所述训练误差,在各个候选子网络中确定最优子网络,并基于所述最优子网络以及主网络确定目标预测模型,能够通过机器学习实现网络模型的自动训练,实现不同网络模型的智能训练和优化,针对不同的广告业务,无需人工进行不同网络模型的训练和优化,节省了大量的人力资源,降低了模型训练与调优的工作量,提高了网络模型的训练效率。The present invention obtains training data and verification data corresponding to preset messages; then generates a plurality of candidate sub-networks through a deep neural sub-network generator, and determines the training error corresponding to each candidate sub-network based on the training data and verification data; Then, based on each of the training errors, the optimal sub-network is determined in each candidate sub-network, and the target prediction model is determined based on the optimal sub-network and the main network, which can realize automatic training of the network model through machine learning, and realize different networks. The intelligent training and optimization of the model does not require manual training and optimization of different network models for different advertising services, which saves a lot of human resources, reduces the workload of model training and tuning, and improves the training efficiency of network models.

附图说明Description of drawings

图1是本发明实施例方案涉及的硬件运行环境中网络模型训练设备的结构示意图;1 is a schematic structural diagram of a network model training device in a hardware operating environment involved in an embodiment of the present invention;

图2为本发明网络模型训练方法第一实施例的流程示意图;2 is a schematic flowchart of a first embodiment of a network model training method according to the present invention;

图3为本发明网络模型训练装置一实施例的功能模块示意图。FIG. 3 is a schematic diagram of functional modules of an embodiment of a network model training apparatus according to the present invention.

本发明目的的实现、功能特点及优点将结合实施例,参照附图做进一步说明。The realization, functional characteristics and advantages of the present invention will be further described with reference to the accompanying drawings in conjunction with the embodiments.

具体实施方式Detailed ways

应当理解,此处所描述的具体实施例仅仅用以解释本发明,并不用于限定本发明。It should be understood that the specific embodiments described herein are only used to explain the present invention, but not to limit the present invention.

如图1所示,图1是本发明实施例方案涉及的硬件运行环境中网络模型训练设备的结构示意图。As shown in FIG. 1 , FIG. 1 is a schematic structural diagram of a network model training device in a hardware operating environment according to an embodiment of the present invention.

本发明实施例终端可以是PC,也可以是智能手机、平板电脑、电子书阅读器、MP3(Moving Picture Experts Group Audio Layer III,动态影像专家压缩标准音频层面3)播放器、MP4(Moving Picture Experts Group Audio Layer IV,动态影像专家压缩标准音频层面4)播放器、便携计算机等具有显示功能的可移动式终端设备。The terminal in the embodiment of the present invention may be a PC, a smart phone, a tablet computer, an e-book reader, an MP3 (Moving Picture Experts Group Audio Layer III, moving picture expert compression standard audio layer 3) player, an MP4 (Moving Picture Experts Group Audio Layer IV, moving image expert compression standard audio layer 4) Players, portable computers and other portable terminal devices with display functions.

如图1所示,该网络模型训练设备可以包括:处理器1001,例如CPU,网络接口1004,用户接口1003,存储器1005,通信总线1002。其中,通信总线1002用于实现这些组件之间的连接通信。用户接口1003可以包括显示屏(Display)、输入单元比如键盘(Keyboard),可选用户接口1003还可以包括标准的有线接口、无线接口。网络接口1004可选的可以包括标准的有线接口、无线接口(如WI-FI接口)。存储器1005可以是高速RAM存储器,也可以是稳定的存储器(non-volatile memory),例如磁盘存储器。存储器1005可选的还可以是独立于前述处理器1001的存储装置。As shown in FIG. 1 , the network model training device may include: a processor 1001 , such as a CPU, a network interface 1004 , a user interface 1003 , a memory 1005 , and a communication bus 1002 . Among them, the communication bus 1002 is used to realize the connection and communication between these components. The user interface 1003 may include a display screen (Display), an input unit such as a keyboard (Keyboard), and the optional user interface 1003 may also include a standard wired interface and a wireless interface. Optionally, the network interface 1004 may include a standard wired interface and a wireless interface (eg, a WI-FI interface). The memory 1005 may be high-speed RAM memory, or may be non-volatile memory, such as disk memory. Optionally, the memory 1005 may also be a storage device independent of the aforementioned processor 1001 .

可选地,网络模型训练设备还可以包括摄像头、RF(Radio Frequency,射频)电路,传感器、音频电路、WiFi模块等等。其中,传感器比如光传感器、运动传感器以及其他传感器。具体地,光传感器可包括环境光传感器及接近传感器;当然,网络模型训练设备还可配置陀螺仪、气压计、湿度计、温度计、红外线传感器等其他传感器,在此不再赘述。Optionally, the network model training device may further include a camera, an RF (Radio Frequency, radio frequency) circuit, a sensor, an audio circuit, a WiFi module, and the like. Among them, sensors such as light sensors, motion sensors and other sensors. Specifically, the light sensor may include an ambient light sensor and a proximity sensor; of course, the network model training device may also be configured with other sensors such as a gyroscope, a barometer, a hygrometer, a thermometer, and an infrared sensor, which will not be repeated here.

本领域技术人员可以理解,图1中示出的网络模型训练设备结构并不构成对网络模型训练设备的限定,可以包括比图示更多或更少的部件,或者组合某些部件,或者不同的部件布置。Those skilled in the art can understand that the structure of the network model training device shown in FIG. 1 does not constitute a limitation on the network model training device, and may include more or less components than those shown in the figure, or combine some components, or different component layout.

如图1所示,作为一种计算机存储介质的存储器1005中可以包括操作系统、网络通信模块、用户接口模块以及网络模型训练程序。As shown in FIG. 1 , the memory 1005 as a computer storage medium may include an operating system, a network communication module, a user interface module and a network model training program.

在图1所示的网络模型训练设备中,网络接口1004主要用于连接后台服务器,与后台服务器进行数据通信;用户接口1003主要用于连接客户端(用户端),与客户端进行数据通信;而处理器1001可以用于调用存储器1005中存储的网络模型训练程序。In the network model training device shown in FIG. 1, the network interface 1004 is mainly used to connect the background server, and perform data communication with the background server; the user interface 1003 is mainly used to connect the client (client), and perform data communication with the client; And the processor 1001 can be used to call the network model training program stored in the memory 1005 .

在本实施例中,网络模型训练设备包括:存储器1005、处理器1001及存储在所述存储器1005上并可在所述处理器1001上运行的网络模型训练程序,其中,处理器1001调用存储器1005中存储的网络模型训练程序时,并执行以下各个实施例中网络模型训练方法的步骤。In this embodiment, the network model training device includes: a memory 1005, a processor 1001, and a network model training program stored on the memory 1005 and running on the processor 1001, wherein the processor 1001 calls the memory 1005 When the network model training program stored in the network model is performed, the steps of the network model training method in the following embodiments are executed.

本发明还提供一种网络模型训练方法,参照图2,图2为本发明网络模型训练方法第一实施例的流程示意图。The present invention also provides a network model training method. Referring to FIG. 2 , FIG. 2 is a schematic flowchart of the first embodiment of the network model training method of the present invention.

本实施例中,该网络模型训练方法包括以下步骤:In this embodiment, the network model training method includes the following steps:

步骤S100,获取预设消息对应的训练数据以及验证数据;Step S100, acquiring training data and verification data corresponding to the preset message;

本实施例中,预设消息可以为预设广告、预设商品、调查问卷等,验证数据为预设消息的点击数据、曝光数据、观看数据等,训练数据为点击数据、曝光数据或观看数据对应的用户特征信息。在对该预设消息进行模型训练时,先获取预设消息对应的训练数据以及验证数据,以将获取到的训练数据以及验证数据作为训练数据,即后续进行模型训练的样本数据。其中,用户特征数据可以包括每一个点击该预设消息的用户的性别、年龄、职业、爱好(例如用户点击频率较高的几个网络链接、或者用户点击频率最高的网络链接)等。In this embodiment, the preset message may be preset advertisements, preset products, questionnaires, etc., the verification data is click data, exposure data, viewing data, etc. of the preset message, and the training data is click data, exposure data, or viewing data Corresponding user feature information. When performing model training on the preset message, first obtain training data and verification data corresponding to the preset message, and use the obtained training data and verification data as training data, that is, sample data for subsequent model training. The user characteristic data may include the gender, age, occupation, and hobby of each user who clicks the preset message (eg, several network links that the user clicks frequently, or the network link that the user clicks the most frequently).

例如,以预设广告为例,该预设广告可以为在线上(网络等)进行投放的任一广告,在该预设广告投放一段时间后,即存在该预设广告对应的曝光点击数据时,可针对该预设广告进行模型训练,得到该预设广告的目标预测模型,例如,可在该预设广告投放之后的持续时间达到预设时长时,对该预设广告进行模型训练,或者,在投放该预设广告之后,检测到预设广告对应的曝光点击数据大于预设曝光点击值时,对该预设广告进行模型训练。For example, taking a preset advertisement as an example, the preset advertisement can be any advertisement placed online (on the Internet, etc.), and after the preset advertisement is placed for a period of time, that is, when there is exposure click data corresponding to the preset advertisement , model training can be performed on the preset advertisement to obtain the target prediction model of the preset advertisement. For example, when the duration after the preset advertisement is placed reaches the preset time length, model training can be performed on the preset advertisement, or , after the preset advertisement is placed, when it is detected that the exposure click data corresponding to the preset advertisement is greater than the preset exposure click value, model training is performed on the preset advertisement.

步骤S200,通过深度神经子网络生成器生成多个候选子网络,并基于所述训练数据以及验证数据,确定各个候选子网络对应的训练误差;Step S200, generating a plurality of candidate sub-networks through a deep neural sub-network generator, and determining the training error corresponding to each candidate sub-network based on the training data and the verification data;

本实施例中,在获取到预设消息对应的训练数据以及验证数据之后,通过深度神经子网络生成器生成多个候选子网络,其中,多个候选子网络中各个子网络的层数不进行限定,即各个候选子网络中的网络层数均为随机的,并且,深度神经子网络生成器每一次生成的候选子网络的数量也是随机的,当然,在其他实施例中,也可以限定候选子网络的数量,例如,每次通过深度神经子网络生成器生成预设个数的候选子网络。In this embodiment, after the training data and verification data corresponding to the preset message are obtained, multiple candidate subnetworks are generated by the deep neural subnetworks generator, wherein the number of layers of each subnetworks in the multiple candidate subnetworks is not calculated. Restriction, that is, the number of network layers in each candidate sub-network is random, and the number of candidate sub-networks generated each time by the deep neural sub-network generator is also random. Of course, in other embodiments, the candidate sub-network can also be limited. The number of sub-networks, for example, a preset number of candidate sub-networks are generated by the deep neural sub-network generator each time.

而后,基于训练数据以及验证数据,确定各个候选子网络对应的训练误差,具体的,将各个候选子网络依次嵌入主网络,并将训练数据以及验证数据输入嵌入候选子网络的主网络,进行模型训练,得到各个候选子网络对应的训练误差。Then, based on the training data and the verification data, the training error corresponding to each candidate sub-network is determined. Specifically, each candidate sub-network is embedded in the main network in turn, and the training data and the verification data are input into the main network embedded in the candidate sub-network, and the model is carried out. Training to get the training error corresponding to each candidate sub-network.

需要说明的是,深度神经子网络生成器可以由算法开发人员进行指定,或者,预先设置多个深度神经子网络生成器,在多个深度神经子网络生成器中随机选取一个用以生成多个候选子网络。It should be noted that the deep neural sub-network generator can be specified by the algorithm developer, or, multiple deep neural sub-network generators are preset, and one of the multiple deep neural sub-network generators is randomly selected to generate multiple candidate subnet.

步骤S300,基于各个所述训练误差,在各个候选子网络中确定最优子网络,并基于所述最优子网络以及主网络确定目标预测模型。Step S300, based on each of the training errors, determine an optimal sub-network in each candidate sub-network, and determine a target prediction model based on the optimal sub-network and the main network.

本实施例中,在获取到训练误差时,基于训练误差,在各个候选子网络中确定最优子网络,例如,将训练误差最小的候选子网络作为最优子网络,或者,将训练误差与各个候选子网络的层数进行权重计算,将权重计算后权重分值最小的候选子网络作为最优子网络。In this embodiment, when the training error is obtained, the optimal sub-network is determined in each candidate sub-network based on the training error. For example, the candidate sub-network with the smallest training error is taken as the optimal sub-network, or the training error is compared with the optimal sub-network. The number of layers of each candidate sub-network is weighted, and the candidate sub-network with the smallest weight score after the weight calculation is used as the optimal sub-network.

而后,基于最优子网络以及主网络确定目标预测模型,其中,若仅需要一次搜索/训练,则将最优子网络嵌入主网络得到的神经网络结构为该目标预测模型;若需要多次搜索/训练,则判断搜索/训练的次数是否达到规定的次数,若是,则将最优子网络嵌入主网络得到的神经网络结构为该目标预测模型,若否,则将最优子网络嵌入主网络,并将该嵌入最优子网络后的主网络作为新的主网络,并返回执行步骤S200,进而能够通过机器学习实现网络模型的自动训练,针对不同的预设消息,无需人工进行不同网络模型的训练和优化,只需要花费算法工程师少量时间就能得到用于线上预设消息投放的机器学习模型,节省了大量的人力资源,提高了网络模型的训练效率。Then, the target prediction model is determined based on the optimal sub-network and the main network. If only one search/training is required, the neural network structure obtained by embedding the optimal sub-network into the main network is the target prediction model; if multiple searches are required /training, then judge whether the number of searches/training has reached the specified number of times, if so, the neural network structure obtained by embedding the optimal sub-network into the main network is the target prediction model, if not, embedding the optimal sub-network into the main network , and use the main network embedded in the optimal sub-network as the new main network, and return to step S200, so that the automatic training of the network model can be realized through machine learning. For different preset messages, there is no need to manually perform different network models. It only takes a small amount of time for algorithm engineers to obtain the machine learning model for online preset message delivery, which saves a lot of human resources and improves the training efficiency of the network model.

进一步地,在一实施例中,步骤S300之后还包括:Further, in an embodiment, after step S300, it further includes:

在确定目标预测模型之后的持续时长达到预设持续时长时,返回执行获取预设消息对应的训练数据以及验证数据的步骤。When the duration after the target prediction model is determined reaches the preset duration, the steps of obtaining training data and verification data corresponding to the preset message are returned to.

本实施例中,通过确定目标预测模型之后的持续时长达到预设持续时长时,返回执行步骤S100,以实现目标预测模型的定时更新,进而基于自动机器学习技术实现目标预测模型的循环更新,可以自动随着用户喜好(用户特征数据)的变化进行模型优化,以保障预设消息的投放效果。In this embodiment, when it is determined that the duration of the target prediction model reaches the preset duration, the process returns to step S100, so as to realize the regular update of the target prediction model, and then realize the cyclic update of the target prediction model based on the automatic machine learning technology. Model optimization is automatically carried out with changes in user preferences (user characteristic data) to ensure the delivery effect of preset messages.

需要说明的是,在其他实施例中,可将预设消息对应的验证数据作为第一验证数据数据进行存储,并实时统计预设消息对应的点击数据、曝光数据或观看数据等,得到当前的第二验证数据,而后计算第二验证数据与第一验证数据之间的差值,在该差值大于预设差值时,返回执行步骤S100,以实现根据用户的点击数量(验证数据)更新目标预测模型,以保障预设消息的投放效果。It should be noted that, in other embodiments, the verification data corresponding to the preset message can be stored as the first verification data data, and the click data, exposure data or viewing data corresponding to the preset message can be counted in real time to obtain the current data. the second verification data, and then calculate the difference between the second verification data and the first verification data, when the difference is greater than the preset difference, return to step S100 to update according to the number of clicks (verification data) of the user Target prediction model to ensure the delivery effect of preset messages.

本实施例提出的网络模型训练方法,通过获取预设消息对应的训练数据以及验证数据;接着通过深度神经子网络生成器生成多个候选子网络,并基于所述训练数据以及验证数据,确定各个候选子网络对应的训练误差;而后基于各个所述训练误差,在各个候选子网络中确定最优子网络,并基于所述最优子网络以及主网络确定目标预测模型,能够通过机器学习实现网络模型的自动训练,实现不同网络模型的智能训练和优化,针对不同的消息业务,无需人工进行不同网络模型的训练和优化,节省了大量的人力资源,降低了模型训练与调优的工作量,提高了网络模型的训练效率。The network model training method proposed in this embodiment obtains training data and verification data corresponding to preset messages; then generates a plurality of candidate sub-networks through a deep neural sub-network generator, and determines each sub-network based on the training data and verification data The training error corresponding to the candidate sub-network; then, based on each of the training errors, the optimal sub-network is determined in each candidate sub-network, and the target prediction model is determined based on the optimal sub-network and the main network, and the network can be realized through machine learning. The automatic training of the model realizes the intelligent training and optimization of different network models. For different message services, there is no need to manually train and optimize different network models, which saves a lot of human resources and reduces the workload of model training and tuning. The training efficiency of the network model is improved.

基于第一实施例,提出本发明网络模型训练方法的第二实施例,在本实施例中,步骤S200包括:Based on the first embodiment, a second embodiment of the network model training method of the present invention is proposed. In this embodiment, step S200 includes:

步骤S210,依次在各个候选子网络获取目标候选子网络;Step S210, obtaining target candidate sub-networks in each candidate sub-network in turn;

步骤S220,将所述目标候选子网络嵌入主网络得到更新后的主网络,并将所述训练数据以及验证数据输入更新后的主网络,以获得所述目标候选子网络对应的训练误差。Step S220: Embed the target candidate sub-network into the main network to obtain an updated main network, and input the training data and verification data into the updated main network to obtain the training error corresponding to the target candidate sub-network.

本实施例中,在得到多个候选子网络之后,依次在各个候选子网络获取目标候选子网络,并将每一次得到的将目标候选子网络嵌入主网络得到更新后的主网络,并将所述训练数据以及验证数据输入更新后的主网络,进行模型训练得到目标候选子网络对应的训练误差,进而准确得到各个候选子网络对应的训练误差。In this embodiment, after a plurality of candidate sub-networks are obtained, target candidate sub-networks are obtained from each candidate sub-network in turn, and each obtained target candidate sub-network is embedded in the main network to obtain an updated main network, and all the obtained target candidate sub-networks are embedded in the main network. The training data and verification data are input into the updated main network, and the training error corresponding to the target candidate sub-network is obtained by model training, and then the training error corresponding to each candidate sub-network is accurately obtained.

需要说明的是,在将所述训练数据以及验证数据输入更新后的主网络,进行模型训练时,可进行多次模型训练,具体的,将所述训练数据以及验证数据输入更新后的主网络,并更新训练次数;若训练次数小于预设训练次数,则再次将所述训练数据以及验证数据输入更新后的主网络,更新后的主网络为已通过训练数据以及验证数据训练过的嵌入目标候选子网络的主网络,若训练次数达到预设训练次数,则将当前更新后的主网络的损失函数值作为训练误差,其中,预设训练次数可进行合理设置。It should be noted that when the training data and verification data are input into the updated main network, and model training is performed, multiple model trainings can be performed. Specifically, the training data and verification data are input into the updated main network. , and update the training times; if the training times are less than the preset training times, input the training data and the verification data into the updated main network again, and the updated main network is the embedded target that has passed the training data and the verification data. For the main network of the candidate sub-network, if the number of training times reaches the preset number of training times, the currently updated loss function value of the main network is used as the training error, where the preset number of training times can be set reasonably.

本实施例中,通过更新后的主网络进行训练后,能够得到训练数据对应的每一个用户点击/查看该预设消息的点击概率,将点击概率小于预设点击概率的用户作为不会点击/查看该预设消息的用户,将不会点击该预设消息的用户除以验证数据的数量即得到训练误差,其中,验证数据的数量可以为训练数据的用户数量。In this embodiment, after training through the updated main network, the click probability of each user corresponding to the training data for clicking/viewing the preset message can be obtained, and the user whose click probability is less than the preset click probability is regarded as not clicking/viewing the preset message. Users who view the preset message will divide the users who do not click the preset message by the number of verification data to obtain the training error, where the number of verification data may be the number of users of the training data.

本实施例提出的网络模型训练方法,通过依次在各个候选子网络获取目标候选子网络;接着将所述目标候选子网络嵌入主网络得到更新后的主网络,并将所述训练数据以及验证数据输入更新后的主网络,以获得所述目标候选子网络对应的训练误差,能够准确得到各个候选子网络对应的训练误差,进而提高了目标预测模型预测该预设消息的准确性,以保障预设消息的投放效果,进一步提高了网络模型的训练效率。In the network model training method proposed in this embodiment, target candidate sub-networks are obtained from each candidate sub-network in turn; then the target candidate sub-network is embedded in the main network to obtain an updated main network, and the training data and verification data are The updated main network is input to obtain the training error corresponding to the target candidate sub-network, and the training error corresponding to each candidate sub-network can be accurately obtained, thereby improving the accuracy of the target prediction model in predicting the preset message, so as to ensure the prediction of the target candidate sub-network. Setting the delivery effect of the message further improves the training efficiency of the network model.

基于第二实施例,提出本发明网络模型训练方法的第三实施例,在本实施例中,步骤S220包括:Based on the second embodiment, a third embodiment of the network model training method of the present invention is proposed. In this embodiment, step S220 includes:

步骤S221,将所述训练数据以及验证数据输入更新后的主网络,并更新训练次数;Step S221, input the training data and verification data into the updated main network, and update the training times;

步骤S222,在所述训练次数达到预设训练次数时,将训练后的更新后的主网络,对应的训练误差作为所述目标候选子网络对应的训练误差;Step S222, when the training times reaches the preset training times, take the training error corresponding to the updated main network after training as the training error corresponding to the target candidate sub-network;

步骤S223,在所述训练次数未达到预设训练次数时,将训练后的更新后的主网络作为更新后的主网络,并返回将所训练数据以及验证数据输入更新后的主网络,并更新训练次数的步骤。Step S223, when the number of times of training does not reach the preset number of times of training, take the updated main network after training as the updated main network, and return the trained data and verification data into the updated main network, and update The number of training steps.

本实施例中,为提高模型训练的效率,需要对更新后的主网络进行多次训练,进而在将所述训练数据以及验证数据输入更新后的主网络,进行更新后的主网络的训练时,更新训练次数即当前更新后的主网络的训练次数。In this embodiment, in order to improve the efficiency of model training, the updated main network needs to be trained multiple times, and then the training data and the verification data are input into the updated main network, and the training of the updated main network is performed. , the number of update training is the number of training of the main network after the current update.

而后,判断训练次数是否达到预设训练次数,若训练次数达到预设训练次数,将训练后的更新后的主网络,对应的训练误差作为所述目标候选子网络对应的训练误差,并重置训练次数,否则将训练后的更新后的主网络作为更新后的主网络,并返回将所述训练数据以及验证数据输入更新后的主网络,并更新训练次数的步骤,进而使得每个目标候选子网络所对应的更新后的主网络,均进行预设训练次数次的训练,提高训练误差的准确性。Then, it is judged whether the training times reaches the preset training times, and if the training times reaches the preset training times, the training error corresponding to the updated main network after training is taken as the training error corresponding to the target candidate sub-network, and reset The number of training times, otherwise the updated main network after training is used as the updated main network, and the training data and verification data are input into the updated main network, and the number of training steps is updated, so that each target candidate The updated main network corresponding to the sub-network is trained for preset training times to improve the accuracy of the training error.

其中,该预设训练次数可进行合理设置。The preset training times can be set reasonably.

本实施例提出的网络模型训练方法,通过将所述训练数据以及验证数据输入更新后的主网络,并更新训练次数;接着在所述训练次数达到预设训练次数时,将训练后的更新后的主网络,对应的训练误差作为所述目标候选子网络对应的训练误差;而后在所述训练次数未达到预设训练次数时,将训练后的更新后的主网络作为更新后的主网络,并返回将所述训练数据以及验证数据输入更新后的主网络,并更新训练次数的步骤,提高对更新后的主网络进行预设训练次数次的训练,进而提高各个候选子网络对应的训练误差的准确性,进一步提高了目标预测模型预测该预设消息的准确性,提高了网络模型的训练效率。In the network model training method proposed in this embodiment, the training data and the verification data are input into the updated main network, and the training times are updated; then when the training times reach the preset training times, the updated the main network, the corresponding training error is taken as the training error corresponding to the target candidate sub-network; then when the number of training times does not reach the preset number of training times, the updated main network after training is used as the updated main network, And return to the steps of inputting the training data and verification data into the updated main network, and updating the training times, so as to improve the training of the updated main network with the preset training times, thereby improving the training error corresponding to each candidate sub-network. The accuracy of the target prediction model further improves the accuracy of predicting the preset message and improves the training efficiency of the network model.

基于第一实施例,提出本发明网络模型训练方法的第四实施例,在本实施例中,步骤S300包括:Based on the first embodiment, a fourth embodiment of the network model training method of the present invention is proposed. In this embodiment, step S300 includes:

步骤S310,获取各个候选子网络的网络层数对应的惩罚系数;Step S310, obtaining the penalty coefficient corresponding to the number of network layers of each candidate sub-network;

步骤S320,基于各个惩罚系数以及训练误差,分别确定各个候选子网络对应的权重分值;Step S320, based on each penalty coefficient and training error, respectively determine the weight score corresponding to each candidate sub-network;

步骤S330,将各个候选子网络中权重分值最小的候选子网络作为所述最优子网络。Step S330, taking the candidate sub-network with the smallest weight score among the candidate sub-networks as the optimal sub-network.

本实施例中,可预先设置网络层数对应的预设惩罚系数,或者网络层数与预设惩罚系数的映射关系,在得到各个候选子网络对应的训练误差之后,根据各个候选子网络的网络层数,以及映射关系,确定各个候选子网络的网络层数对应的惩罚系数。In this embodiment, the preset penalty coefficient corresponding to the number of network layers, or the mapping relationship between the number of network layers and the preset penalty coefficient can be preset. The number of layers, and the mapping relationship, determine the penalty coefficient corresponding to the number of network layers of each candidate sub-network.

而后,基于各个惩罚系数以及训练误差,分别确定各个候选子网络对应的权重分值,具体的,分别获取惩罚系数的权重以及训练误差的权重,基于惩罚系数以及训练误差、惩罚系数的权重以及训练误差的权重确定各个候选子网络对应的权重分值,例如,权重分值=惩罚系数的权重*惩罚系数+训练误差的权重*训练误差。Then, based on each penalty coefficient and training error, the corresponding weight scores of each candidate sub-network are determined respectively. Specifically, the weight of the penalty coefficient and the weight of the training error are obtained respectively, based on the penalty coefficient and the training error, the weight of the penalty coefficient and the training error. The weight of the error determines the weight score corresponding to each candidate sub-network, for example, weight score=weight of penalty coefficient*penalty coefficient+weight of training error*training error.

本实施例提出的网络模型训练方法,通过获取各个候选子网络的网络层数对应的惩罚系数;接着基于各个惩罚系数以及训练误差,分别确定各个候选子网络对应的权重分值;而后将各个候选子网络中权重分值最小的候选子网络作为所述最优子网络,通过综合考量惩罚系数以及训练误差,使得确定的候选子网络为网络简单且训练误差较小的候选子网络,在提高了目标预测模型预测该预设消息的准确性的同时,降低网络模型的复杂性。The network model training method proposed in this embodiment obtains the penalty coefficient corresponding to the number of network layers of each candidate sub-network; then, based on each penalty coefficient and the training error, respectively determines the weight score corresponding to each candidate sub-network; The candidate sub-network with the smallest weight score in the sub-network is used as the optimal sub-network. By comprehensively considering the penalty coefficient and the training error, the determined candidate sub-network is a candidate sub-network with a simple network and a small training error. The target prediction model reduces the complexity of the network model while predicting the accuracy of the preset message.

基于第一实施例,提出本发明网络模型训练方法的第五实施例,在本实施例中,该网络模型训练方法还包括:Based on the first embodiment, a fifth embodiment of the network model training method of the present invention is proposed. In this embodiment, the network model training method further includes:

步骤S100,在通过深度神经子网络生成器生成多个候选子网络时,更新搜索次数;Step S100, when multiple candidate sub-networks are generated by the deep neural sub-network generator, update the number of searches;

步骤S300包括:Step S300 includes:

步骤S340,确定所述最优子网络对应的训练误差是否小于所述主网络对应的训练误差;Step S340, determining whether the training error corresponding to the optimal sub-network is smaller than the training error corresponding to the main network;

步骤S350,若所述最优子网络对应的训练误差小于所述主网络对应的训练误差,则确定所述搜索次数是否达到预设搜索次数;Step S350, if the training error corresponding to the optimal sub-network is smaller than the training error corresponding to the main network, determine whether the number of searches reaches a preset number of searches;

步骤S360,在所述搜索次数达到所述预设搜索次数时,则将所述最优子网络嵌入所述主网络以获得所述目标预测模型。Step S360, when the number of searches reaches the preset number of searches, the optimal sub-network is embedded in the main network to obtain the target prediction model.

本实施例中,为提高目标预测模型的预测准确性,需要叠加多个最优子网络,因此,在通过深度神经子网络生成器生成多个候选子网络时,更新搜索次数,具体的,将搜索次数+1得到更新后的搜索次数。In this embodiment, in order to improve the prediction accuracy of the target prediction model, multiple optimal sub-networks need to be superimposed. Therefore, when multiple candidate sub-networks are generated by the deep neural sub-network generator, the number of searches is updated. Search count +1 to get the updated search count.

而后,确定所述最优子网络对应的训练误差是否小于所述主网络对应的训练误差,其中,在主网络中为嵌入任一子网络时,该主网络的训练误差大于任一候选子网络对应的训练误差,在主网络中已嵌入子网络时,可根据前一次的训练数据得到主网络的训练误差。Then, it is determined whether the training error corresponding to the optimal sub-network is smaller than the training error corresponding to the main network, wherein, when any sub-network is embedded in the main network, the training error of the main network is greater than that of any candidate sub-network For the corresponding training error, when the sub-network has been embedded in the main network, the training error of the main network can be obtained according to the previous training data.

若最优子网络对应的训练误差小于主网络对应的训练误差,则确定所述搜索次数是否达到预设搜索次数,以判断当前是否已完成预设搜索次数的搜索/训练,在所述搜索次数达到所述预设搜索次数时,则将所述最优子网络嵌入所述主网络以获得所述目标预测模型,并同时重置搜索次数。If the training error corresponding to the optimal sub-network is smaller than the training error corresponding to the main network, it is determined whether the number of searches reaches the preset number of searches to determine whether the search/training of the preset number of searches has been completed currently. When the preset number of searches is reached, the optimal sub-network is embedded in the main network to obtain the target prediction model, and the number of searches is reset at the same time.

进一步地,在一实施例中,步骤S340之后,还包括:Further, in an embodiment, after step S340, it further includes:

若所述最优子网络对应的训练误差大于或等于所述主网络对应的训练误差,则返回执行通过深度神经子网络生成器生成多个候选子网络的步骤。If the training error corresponding to the optimal sub-network is greater than or equal to the training error corresponding to the main network, return to the step of generating multiple candidate sub-networks through the deep neural sub-network generator.

本实施例中,若最优子网络对应的训练误差大于或等于所述主网络对应的训练误差,则放弃对当前的主网络进行更新,并返回执行通过深度神经子网络生成器生成多个候选子网络的步骤,其中,在本次生成多个候选子网络时,可更新或者不更新搜索次数;或者,若此时,搜索次数达到所述预设搜索次数,则本次生成多个候选子网络时,不更新搜索次数,若搜索次数未达到所述预设搜索次数,则更新搜索次数。In this embodiment, if the training error corresponding to the optimal sub-network is greater than or equal to the training error corresponding to the main network, the update of the current main network is abandoned, and the execution of generating multiple candidates through the deep neural sub-network generator is returned. The step of sub-network, wherein, when multiple candidate sub-networks are generated this time, the number of searches may or may not be updated; or, if the number of searches reaches the preset number of searches at this time, multiple candidate sub-networks are generated this time. In the network, the number of searches is not updated, and if the number of searches does not reach the preset number of searches, the number of searches is updated.

进一步地,又一实施例中,步骤S350之后,还包括:Further, in another embodiment, after step S350, it further includes:

步骤S360,在所述搜索次数未达到所述预设搜索次数时,则将所述最优子网络嵌入所述主网络;Step S360, when the number of searches does not reach the preset number of searches, embed the optimal sub-network into the main network;

步骤S370,将嵌入所述最优子网络后的主网络作为所述主网络,并返回执行通过深度神经子网络生成器生成多个候选子网络的步骤。Step S370, take the main network after embedding the optimal sub-network as the main network, and return to the step of generating multiple candidate sub-networks through the deep neural sub-network generator.

本实施例中,搜索次数未达到所述预设搜索次数,模型训练未达到规定的次数,因此,将所述最优子网络嵌入所述主网络,将嵌入所述最优子网络后的主网络作为所述主网络,并返回执行通过深度神经子网络生成器生成多个候选子网络的步骤,以实现循环训练。In this embodiment, the number of searches has not reached the preset number of searches, and the model training has not reached the specified number of times. Therefore, the optimal sub-network is embedded in the main network, and the main network after embedding the optimal sub-network is The network acts as the main network, and returns to perform the step of generating multiple candidate sub-networks through the deep neural sub-network generator to realize cyclic training.

本实施例提出的网络模型训练方法,通过确定所述最优子网络对应的训练误差是否小于所述主网络对应的训练误差;接着若所述最优子网络对应的训练误差小于所述主网络对应的训练误差,则确定所述搜索次数是否达到预设搜索次数;而后在所述搜索次数达到所述预设搜索次数时,则将所述最优子网络嵌入所述主网络以获得所述目标预测模型,通过多次搜索得到目标预测模型,进一步提高了目标预测模型预测该预设消息的准确性,提升模型的训练效率。The network model training method proposed in this embodiment determines whether the training error corresponding to the optimal sub-network is smaller than the training error corresponding to the main network; then if the training error corresponding to the optimal sub-network is smaller than the main network corresponding training error, then determine whether the number of searches reaches the preset number of searches; then when the number of searches reaches the preset number of searches, the optimal sub-network is embedded in the main network to obtain the For the target prediction model, the target prediction model is obtained through multiple searches, which further improves the accuracy of the target prediction model in predicting the preset message and improves the training efficiency of the model.

基于上述各个实施例,提出本发明网络模型训练方法的第六实施例,在本实施例中,所述训练数据为用户特征数据,所述验证数据为曝光点击数据,步骤S300之后,该网络模型训练方法还包括:Based on the above embodiments, a sixth embodiment of the network model training method of the present invention is proposed. In this embodiment, the training data is user characteristic data, and the verification data is exposure click data. After step S300, the network model Training methods also include:

步骤S400,获取待预测用户特征数据,并将所述待预测用户数据输入所述目标预测模型,以获得预设消息对应的点击概率;Step S400, obtaining the feature data of the user to be predicted, and inputting the user data to be predicted into the target prediction model to obtain the click probability corresponding to the preset message;

步骤S500,在所述点击概率大于预设阈值时,确定所述待预测用户特征数据对应的用户为待投放用户。Step S500, when the click probability is greater than a preset threshold, determine that the user corresponding to the feature data of the user to be predicted is the user to be placed.

本实施例中,在得到目标预测模型之后,可随时获取待预测用户特征数据,并将待预测用户数据输入所述目标预测模型,进行模型训练得到得预设消息对应的点击概率,即该待预测用户特征数据点击该预设消息的点击概率。In this embodiment, after the target prediction model is obtained, the characteristic data of the user to be predicted can be obtained at any time, and the user data to be predicted can be input into the target prediction model, and the click probability corresponding to the preset message can be obtained through model training, that is, the click probability corresponding to the preset message is obtained. Predict the click probability that the user characteristic data clicks the preset message.

而后,判断该点击概率是否大于预设阈值,若大于则确定所述待预测用户特征数据对应的用户为待投放用户,可将预设消息投放至所述待预测用户特征数据对应的用户终端,或者,将待投放用户的用户终端信息推送至预设消息的投放服务器,以使投放服务器将预设消息投放至所述待预测用户特征数据对应的用户终端,以实现预设消息的准确投放,确保投放的预设消息能够大概率被用户点击查看,进而保障预设消息的投放效果。Then, it is judged whether the click probability is greater than a preset threshold, and if it is greater than the value, it is determined that the user corresponding to the feature data of the user to be predicted is the user to be delivered, and the preset message can be delivered to the user terminal corresponding to the feature data of the user to be predicted. Or, push the user terminal information of the user to be delivered to the delivery server of the preset message, so that the delivery server will deliver the preset message to the user terminal corresponding to the feature data of the user to be predicted, so as to achieve accurate delivery of the preset message, Ensure that the delivered preset messages can be clicked and viewed by users with a high probability, thereby ensuring the delivery effect of the preset messages.

例如,在某一用户刷观看某视频时,可获取该用户的特征数据作为待预测用户特征数据,进行模型训练得到该用户点击该预设消息的概率,若点击概率大于预设阈值,则在用户完成视频观看时推送该预设消息至该用户的终端。For example, when a user watches a certain video, the user's feature data can be obtained as the user's feature data to be predicted, and model training is performed to obtain the probability that the user clicks on the preset message. If the click probability is greater than the preset threshold, then in When the user finishes watching the video, the preset message is pushed to the user's terminal.

本实施例提出的网络模型训练方法,通过获取待预测用户特征数据,并将所述待预测用户数据输入所述目标预测模型进行模型训练,以获得预设消息对应的点击概率;接着在所述点击概率大于预设阈值时,确定所述待预测用户特征数据对应的用户为待投放用户,通过准确预测用户的点击概率进行预设消息投放,提升了预设消息的投放效果。The network model training method proposed in this embodiment obtains the feature data of the user to be predicted, and inputs the user data to be predicted into the target prediction model for model training, so as to obtain the click probability corresponding to the preset message; When the click probability is greater than the preset threshold, it is determined that the user corresponding to the feature data of the user to be predicted is the user to be delivered, and the delivery of the preset message is performed by accurately predicting the click probability of the user, thereby improving the delivery effect of the preset message.

本发明实施例还提供一种网络模型训练装置,参照图3,所述网络模型训练装置包括:An embodiment of the present invention further provides a network model training device. Referring to FIG. 3 , the network model training device includes:

获取模块100,用于获取预设消息对应的训练数据以及验证数据;an acquisition module 100, configured to acquire training data and verification data corresponding to the preset message;

生成模块200,用于通过深度神经子网络生成器生成多个候选子网络,并基于所述训练数据以及验证数据,确定各个候选子网络对应的训练误差;The generating module 200 is used for generating a plurality of candidate sub-networks through a deep neural sub-network generator, and based on the training data and the verification data, determining the training error corresponding to each candidate sub-network;

确定模块300,用于基于各个所述训练误差,在各个候选子网络中确定最优子网络,接着基于所述最优子网络以及主网络确定目标预测模型。The determining module 300 is configured to determine an optimal sub-network in each candidate sub-network based on each of the training errors, and then determine a target prediction model based on the optimal sub-network and the main network.

可选地,生成模块200还用于:Optionally, the generating module 200 is also used for:

依次在各个候选子网络获取目标候选子网络;Obtain target candidate sub-networks in each candidate sub-network in turn;

将所述目标候选子网络嵌入主网络得到更新后的主网络,并将所述训练数据以及验证数据输入更新后的主网络,以获得所述目标候选子网络对应的训练误差。Embed the target candidate sub-network into the main network to obtain an updated main network, and input the training data and verification data into the updated main network to obtain the training error corresponding to the target candidate sub-network.

可选地,生成模块200还用于:Optionally, the generating module 200 is also used for:

将所述训练数据以及验证数据输入更新后的主网络,并更新训练次数;Input the training data and the verification data into the updated main network, and update the training times;

在所述训练次数达到预设训练次数时,将训练后的更新后的主网络,对应的训练误差作为所述目标候选子网络对应的训练误差;When the number of training times reaches the preset number of training times, the training error corresponding to the updated main network after training is used as the training error corresponding to the target candidate sub-network;

在所述训练次数未达到预设训练次数时,将训练后的更新后的主网络作为更新后的主网络,并返回将所述训练数据以及验证数据输入更新后的主网络,并更新训练次数的步骤。When the number of trainings does not reach the preset number of trainings, take the updated main network after training as the updated main network, and return the training data and verification data into the updated main network, and update the training times A step of.

可选地,确定模块300还用于:Optionally, the determining module 300 is further configured to:

获取各个候选子网络的网络层数对应的惩罚系数;Obtain the penalty coefficient corresponding to the number of network layers of each candidate sub-network;

基于各个惩罚系数以及训练误差,分别确定各个候选子网络对应的权重分值;Based on each penalty coefficient and training error, the corresponding weight scores of each candidate sub-network are determined respectively;

将各个候选子网络中权重分值最小的候选子网络作为所述最优子网络。The candidate sub-network with the smallest weight score among the candidate sub-networks is used as the optimal sub-network.

可选地,网络模型训练装置还包括:Optionally, the network model training device further includes:

在通过深度神经子网络生成器生成多个候选子网络时,更新搜索次数;When multiple candidate sub-networks are generated by the deep neural sub-network generator, update the number of searches;

确定模块300还用于:The determination module 300 is also used to:

确定所述最优子网络对应的训练误差是否小于所述主网络对应的训练误差;Determine whether the training error corresponding to the optimal sub-network is smaller than the training error corresponding to the main network;

若所述最优子网络对应的训练误差小于所述主网络对应的训练误差,则确定所述搜索次数是否达到预设搜索次数;If the training error corresponding to the optimal sub-network is smaller than the training error corresponding to the main network, determine whether the number of searches reaches a preset number of searches;

在所述搜索次数达到所述预设搜索次数时,则将所述最优子网络嵌入所述主网络以获得所述目标预测模型。When the number of searches reaches the preset number of searches, the optimal sub-network is embedded in the main network to obtain the target prediction model.

可选地,确定模块300还用于:Optionally, the determining module 300 is further configured to:

若所述最优子网络对应的训练误差大于或等于所述主网络对应的训练误差,则返回执行通过深度神经子网络生成器生成多个候选子网络的步骤。If the training error corresponding to the optimal sub-network is greater than or equal to the training error corresponding to the main network, return to the step of generating multiple candidate sub-networks through the deep neural sub-network generator.

可选地,确定模块300还用于:Optionally, the determining module 300 is further configured to:

在所述搜索次数未达到所述预设搜索次数时,则将所述最优子网络嵌入所述主网络;When the number of searches does not reach the preset number of searches, the optimal sub-network is embedded in the main network;

将嵌入所述最优子网络后的主网络作为所述主网络,并返回执行通过深度神经子网络生成器生成多个候选子网络的步骤。The main network after embedding the optimal sub-network is used as the main network, and the step of generating a plurality of candidate sub-networks by the deep neural sub-network generator is performed back.

可选地,网络模型训练装置还包括:Optionally, the network model training device further includes:

在确定目标预测模型之后的持续时长达到预设持续时长时,返回执行获取预设消息对应的训练数据以及验证数据的步骤。When the duration after the target prediction model is determined reaches the preset duration, the steps of acquiring training data and verification data corresponding to the preset message are returned to.

可选地,网络模型训练装置还包括:Optionally, the network model training device further includes:

获取待预测用户特征数据,并将所述待预测用户数据输入所述目标预测模型,以获得预设消息对应的点击概率;Obtaining the feature data of the user to be predicted, and inputting the user data to be predicted into the target prediction model to obtain the click probability corresponding to the preset message;

在所述点击概率大于预设阈值时,确定所述待预测用户特征数据对应的用户为待投放用户。When the click probability is greater than a preset threshold, it is determined that the user corresponding to the to-be-predicted user characteristic data is the to-be-delivered user.

上述各程序模块所执行的方法可参照本发明网络模型训练方法各个实施例,此处不再赘述。For the methods executed by the above program modules, reference may be made to the various embodiments of the network model training method of the present invention, which will not be repeated here.

此外,本发明实施例还提出一种计算机可读存储介质,所述计算机可读存储介质上存储有网络模型训练程序,所述网络模型训练程序被处理器执行时实现如上所述的网络模型训练方法的步骤。In addition, an embodiment of the present invention also provides a computer-readable storage medium, where a network model training program is stored on the computer-readable storage medium, and when the network model training program is executed by a processor, the above-mentioned network model training is implemented steps of the method.

其中,在所述处理器上运行的网络模型训练程序被执行时所实现的方法可参照本发明网络模型训练方法各个实施例,此处不再赘述。For the method implemented when the network model training program running on the processor is executed, reference may be made to the various embodiments of the network model training method of the present invention, which will not be repeated here.

需要说明的是,在本文中,术语“包括”、“包含”或者其任何其他变体意在涵盖非排他性的包含,从而使得包括一系列要素的过程、方法、物品或者系统不仅包括那些要素,而且还包括没有明确列出的其他要素,或者是还包括为这种过程、方法、物品或者系统所固有的要素。在没有更多限制的情况下,由语句“包括一个……”限定的要素,并不排除在包括该要素的过程、方法、物品或者系统中还存在另外的相同要素。It should be noted that, herein, the terms "comprising", "comprising" or any other variation thereof are intended to encompass non-exclusive inclusion, such that a process, method, article or system comprising a series of elements includes not only those elements, It also includes other elements not expressly listed or inherent to such a process, method, article or system. Without further limitation, an element qualified by the phrase "comprising a..." does not preclude the presence of additional identical elements in the process, method, article or system that includes the element.

上述本发明实施例序号仅仅为了描述,不代表实施例的优劣。The above-mentioned serial numbers of the embodiments of the present invention are only for description, and do not represent the advantages or disadvantages of the embodiments.

通过以上的实施方式的描述,本领域的技术人员可以清楚地了解到上述实施例方法可借助软件加必需的通用硬件平台的方式来实现,当然也可以通过硬件,但很多情况下前者是更佳的实施方式。基于这样的理解,本发明的技术方案本质上或者说对现有技术做出贡献的部分可以以软件产品的形式体现出来,该计算机软件产品存储在如上所述的一个存储介质(如ROM/RAM、磁碟、光盘)中,包括若干指令用以使得一台终端设备(可以是手机,计算机,服务器,空调器,或者网络设备等)执行本发明各个实施例所述的方法。From the description of the above embodiments, those skilled in the art can clearly understand that the method of the above embodiment can be implemented by means of software plus a necessary general hardware platform, and of course can also be implemented by hardware, but in many cases the former is better implementation. Based on such understanding, the technical solutions of the present invention can be embodied in the form of software products in essence or the parts that make contributions to the prior art, and the computer software products are stored in a storage medium (such as ROM/RAM) as described above. , magnetic disk, optical disk), including several instructions to make a terminal device (which may be a mobile phone, a computer, a server, an air conditioner, or a network device, etc.) execute the methods described in the various embodiments of the present invention.

以上仅为本发明的优选实施例,并非因此限制本发明的专利范围,凡是利用本发明说明书及附图内容所作的等效结构或等效流程变换,或直接或间接运用在其他相关的技术领域,均同理包括在本发明的专利保护范围内。The above are only preferred embodiments of the present invention, and are not intended to limit the scope of the present invention. Any equivalent structure or equivalent process transformation made by using the contents of the description and drawings of the present invention, or directly or indirectly applied in other related technical fields , are similarly included in the scope of patent protection of the present invention.

Claims (12)

1.一种网络模型训练方法,其特征在于,所述网络模型训练方法包括以下步骤:1. a network model training method, is characterized in that, described network model training method comprises the following steps: 获取预设消息对应的训练数据以及验证数据;Obtain the training data and verification data corresponding to the preset message; 通过深度神经子网络生成器生成多个候选子网络,并基于所述训练数据以及验证数据,确定各个候选子网络对应的训练误差;Generate a plurality of candidate sub-networks by the deep neural sub-network generator, and determine the training error corresponding to each candidate sub-network based on the training data and the verification data; 基于各个所述训练误差,在各个候选子网络中确定最优子网络,并基于所述最优子网络以及主网络确定目标预测模型。Based on each of the training errors, an optimal sub-network is determined in each candidate sub-network, and a target prediction model is determined based on the optimal sub-network and the main network. 2.如权利要求1所述的网络模型训练方法,其特征在于,所述基于所述训练数据以及验证数据,确定各个候选子网络对应的训练误差的步骤包括:2. The network model training method according to claim 1, wherein the step of determining the training error corresponding to each candidate sub-network based on the training data and the verification data comprises: 依次在各个候选子网络获取目标候选子网络;Obtain target candidate sub-networks in each candidate sub-network in turn; 将所述目标候选子网络嵌入主网络得到更新后的主网络,并将所述训练数据以及验证数据输入更新后的主网络,以获得所述目标候选子网络对应的训练误差。Embed the target candidate sub-network into the main network to obtain an updated main network, and input the training data and verification data into the updated main network to obtain the training error corresponding to the target candidate sub-network. 3.如权利要求2所述的网络模型训练方法,其特征在于,所述将所述训练数据以及验证数据输入更新后的主网络,以获得所述目标候选子网络对应的训练误差的步骤包括:3. The network model training method according to claim 2, wherein the step of inputting the training data and the verification data into the updated main network to obtain the training error corresponding to the target candidate sub-network comprises the following steps: : 将所述训练数据以及验证数据输入更新后的主网络,并更新训练次数;Input the training data and the verification data into the updated main network, and update the training times; 在所述训练次数达到预设训练次数时,将训练后的更新后的主网络,对应的训练误差作为所述目标候选子网络对应的训练误差;When the number of training times reaches the preset number of training times, the training error corresponding to the updated main network after training is used as the training error corresponding to the target candidate sub-network; 在所述训练次数未达到预设训练次数时,将训练后的更新后的主网络作为更新后的主网络,并返回将所述训练数据以及验证数据输入更新后的主网络,并更新训练次数的步骤。When the number of trainings does not reach the preset number of trainings, the updated main network after training is used as the updated main network, and the training data and verification data are input into the updated main network, and the training times are updated. A step of. 4.如权利要求1所述的网络模型训练方法,其特征在于,所述基于各个所述训练误差,在各个候选子网络中确定最优子网络的步骤包括:4. The network model training method according to claim 1, wherein the step of determining an optimal sub-network in each candidate sub-network based on each of the training errors comprises: 获取各个候选子网络的网络层数对应的惩罚系数;Obtain the penalty coefficient corresponding to the number of network layers of each candidate sub-network; 基于各个惩罚系数以及训练误差,分别确定各个候选子网络对应的权重分值;Based on each penalty coefficient and training error, the corresponding weight scores of each candidate sub-network are determined respectively; 将各个候选子网络中权重分值最小的候选子网络作为所述最优子网络。The candidate sub-network with the smallest weight score among the candidate sub-networks is used as the optimal sub-network. 5.如权利要求1所述的网络模型训练方法,其特征在于,所述网络模型训练方法还包括:5. The network model training method according to claim 1, wherein the network model training method further comprises: 在通过深度神经子网络生成器生成多个候选子网络时,更新搜索次数;When multiple candidate sub-networks are generated by the deep neural sub-network generator, update the number of searches; 所述基于所述最优子网络以及主网络确定所述目标预测模型的步骤包括:The step of determining the target prediction model based on the optimal sub-network and the main network includes: 确定所述最优子网络对应的训练误差是否小于所述主网络对应的训练误差;Determine whether the training error corresponding to the optimal sub-network is smaller than the training error corresponding to the main network; 若所述最优子网络对应的训练误差小于所述主网络对应的训练误差,则确定所述搜索次数是否达到预设搜索次数;If the training error corresponding to the optimal sub-network is smaller than the training error corresponding to the main network, determine whether the number of searches reaches a preset number of searches; 在所述搜索次数达到所述预设搜索次数时,则将所述最优子网络嵌入所述主网络以获得所述目标预测模型。When the number of searches reaches the preset number of searches, the optimal sub-network is embedded in the main network to obtain the target prediction model. 6.如权利要求5所述的网络模型训练方法,其特征在于,所述确定所述最优子网络对应的训练误差是否小于所述主网络对应的训练误差的步骤之后,还包括:6. The network model training method according to claim 5, wherein after the step of determining whether the training error corresponding to the optimal sub-network is less than the training error corresponding to the main network, further comprising: 若所述最优子网络对应的训练误差大于或等于所述主网络对应的训练误差,则返回执行通过深度神经子网络生成器生成多个候选子网络的步骤。If the training error corresponding to the optimal sub-network is greater than or equal to the training error corresponding to the main network, return to the step of generating multiple candidate sub-networks through the deep neural sub-network generator. 7.如权利要求5所述的网络模型训练方法,其特征在于,所述确定所述搜索次数是否达到预设搜索次数的步骤之后,还包括:7. The network model training method according to claim 5, wherein after the step of determining whether the number of searches reaches a preset number of searches, further comprising: 在所述搜索次数未达到所述预设搜索次数时,则将所述最优子网络嵌入所述主网络;When the number of searches does not reach the preset number of searches, the optimal sub-network is embedded in the main network; 将嵌入所述最优子网络后的主网络作为所述主网络,并返回执行通过深度神经子网络生成器生成多个候选子网络的步骤。The main network after embedding the optimal sub-network is used as the main network, and the step of generating a plurality of candidate sub-networks by the deep neural sub-network generator is performed back. 8.如权利要求1所述的网络模型训练方法,其特征在于,所述基于所述最优子网络以及主网络确定目标预测模型的步骤之后,所述网络模型训练方法还包括:8. The network model training method according to claim 1, wherein after the step of determining the target prediction model based on the optimal sub-network and the main network, the network model training method further comprises: 在确定目标预测模型之后的持续时长达到预设持续时长时,返回执行获取预设消息对应的训练数据以及验证数据的步骤。When the duration after the target prediction model is determined reaches the preset duration, the steps of acquiring training data and verification data corresponding to the preset message are returned to. 9.如权利要求1至8任一项所述的网络模型训练方法,其特征在于,所述训练数据为用户特征数据,所述验证数据为曝光点击数据,所述基于所述最优子网络以及主网络确定目标预测模型的步骤之后,所述网络模型训练还包括:9. The network model training method according to any one of claims 1 to 8, wherein the training data is user feature data, the verification data is exposure click data, and the optimal sub-network is based on the And after the step of determining the target prediction model by the main network, the network model training further includes: 获取待预测用户特征数据,并将所述待预测用户数据输入所述目标预测模型,以获得预设消息对应的点击概率;Obtaining the feature data of the user to be predicted, and inputting the user data to be predicted into the target prediction model to obtain the click probability corresponding to the preset message; 在所述点击概率大于预设阈值时,确定所述待预测用户特征数据对应的用户为待投放用户。When the click probability is greater than a preset threshold, it is determined that the user corresponding to the to-be-predicted user characteristic data is the to-be-delivered user. 10.一种网络模型训练装置,其特征在于,所述网络模型训练装置包括:10. A network model training device, wherein the network model training device comprises: 获取模块,用于获取预设消息对应的训练数据以及验证数据;an acquisition module for acquiring training data and verification data corresponding to the preset message; 生成模块,用于通过深度神经子网络生成器生成多个候选子网络,并基于所述训练数据以及验证数据,确定各个候选子网络对应的训练误差;A generation module, used for generating multiple candidate sub-networks through the deep neural sub-network generator, and determining the training error corresponding to each candidate sub-network based on the training data and the verification data; 确定模块,用于基于各个所述训练误差,在各个候选子网络中确定最优子网络,接着基于所述最优子网络以及主网络确定目标预测模型。A determination module, configured to determine an optimal sub-network in each candidate sub-network based on each of the training errors, and then determine a target prediction model based on the optimal sub-network and the main network. 11.一种网络模型训练设备,其特征在于,所述网络模型训练设备包括:存储器、处理器及存储在所述存储器上并可在所述处理器上运行的网络模型训练程序,所述网络模型训练程序被所述处理器执行时实现如权利要求1至9中任一项所述的网络模型训练方法的步骤。11. A network model training device, characterized in that the network model training device comprises: a memory, a processor, and a network model training program stored on the memory and running on the processor, the network When the model training program is executed by the processor, the steps of the network model training method according to any one of claims 1 to 9 are implemented. 12.一种计算机可读存储介质,其特征在于,所述计算机可读存储介质上存储有网络模型训练程序,所述网络模型训练程序被处理器执行时实现如权利要求1至9中任一项所述的网络模型训练方法的步骤。12. A computer-readable storage medium, characterized in that, a network model training program is stored on the computer-readable storage medium, and when the network model training program is executed by a processor, any one of claims 1 to 9 is implemented. The steps of the network model training method described in item.
CN202010145962.1A 2020-03-04 2020-03-04 Network model training method, device, equipment and computer readable storage medium Active CN111291868B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010145962.1A CN111291868B (en) 2020-03-04 2020-03-04 Network model training method, device, equipment and computer readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010145962.1A CN111291868B (en) 2020-03-04 2020-03-04 Network model training method, device, equipment and computer readable storage medium

Publications (2)

Publication Number Publication Date
CN111291868A true CN111291868A (en) 2020-06-16
CN111291868B CN111291868B (en) 2025-11-11

Family

ID=71021642

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010145962.1A Active CN111291868B (en) 2020-03-04 2020-03-04 Network model training method, device, equipment and computer readable storage medium

Country Status (1)

Country Link
CN (1) CN111291868B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112734005A (en) * 2020-12-31 2021-04-30 北京达佳互联信息技术有限公司 Method and device for determining prediction model, electronic equipment and storage medium
CN113986770A (en) * 2021-12-27 2022-01-28 深圳市明源云科技有限公司 User system upgrade method, device, equipment and medium based on artificial intelligence

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107025205A (en) * 2016-01-30 2017-08-08 华为技术有限公司 A kind of method and apparatus of training pattern in distributed system
CN107403222A (en) * 2017-07-19 2017-11-28 燕山大学 A kind of motion tracking method based on auxiliary more new model and validity check
CN110782034A (en) * 2019-10-31 2020-02-11 北京小米智能科技有限公司 Neural network training method, device and storage medium

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107025205A (en) * 2016-01-30 2017-08-08 华为技术有限公司 A kind of method and apparatus of training pattern in distributed system
CN107403222A (en) * 2017-07-19 2017-11-28 燕山大学 A kind of motion tracking method based on auxiliary more new model and validity check
CN110782034A (en) * 2019-10-31 2020-02-11 北京小米智能科技有限公司 Neural network training method, device and storage medium

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112734005A (en) * 2020-12-31 2021-04-30 北京达佳互联信息技术有限公司 Method and device for determining prediction model, electronic equipment and storage medium
CN113986770A (en) * 2021-12-27 2022-01-28 深圳市明源云科技有限公司 User system upgrade method, device, equipment and medium based on artificial intelligence
CN113986770B (en) * 2021-12-27 2022-04-22 深圳市明源云科技有限公司 User system upgrading method, device, equipment and medium based on artificial intelligence

Also Published As

Publication number Publication date
CN111291868B (en) 2025-11-11

Similar Documents

Publication Publication Date Title
CN109902708B (en) A recommendation model training method and related device
CN111242310B (en) Feature validity evaluation method and device, electronic equipment and storage medium
US10965775B2 (en) Discovering signature of electronic social networks
CN108280115B (en) Method and device for identifying user relationship
CN109086439A (en) Information recommendation method and device
CN111783810B (en) Method and device for determining user attribute information
CN114119123A (en) Information pushing method and device
CN111538907A (en) Object recommendation method, system and device
CN107862053A (en) User's portrait building method, device and computing device based on customer relationship
US20190156231A1 (en) User segmentation using predictive model interpretation
WO2023050143A1 (en) Recommendation model training method and apparatus
WO2024002167A1 (en) Operation prediction method and related apparatus
CN110837577A (en) Video recommendation method, device, equipment and storage medium
CN112100558B (en) Method, device, apparatus and storage medium for object recommendation
CN116204709A (en) Data processing method and related device
CN114430504B (en) Recommendation method and related device for media content
CN114238766B (en) Material recommendation method, material recommendation device and electronic equipment
CN113836388A (en) Information recommendation method and device, server and storage medium
CN116401400B (en) Model training methods and related equipment
CN111291868B (en) Network model training method, device, equipment and computer readable storage medium
US20240386315A1 (en) Transformer model for journey simulation and prediction
WO2024152686A1 (en) Method and apparatus for determining recommendation index of resource information, device, storage medium and computer program product
CN111507471B (en) A model training method, device, equipment and storage medium
CN111784091B (en) Method and apparatus for processing information
CN118779530A (en) Information recommendation method, device, storage medium and electronic device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant