[go: up one dir, main page]

CN113919506B - A model training online method, device, storage medium and equipment - Google Patents

A model training online method, device, storage medium and equipment Download PDF

Info

Publication number
CN113919506B
CN113919506B CN202111175935.XA CN202111175935A CN113919506B CN 113919506 B CN113919506 B CN 113919506B CN 202111175935 A CN202111175935 A CN 202111175935A CN 113919506 B CN113919506 B CN 113919506B
Authority
CN
China
Prior art keywords
model
test
online
evaluation value
training
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202111175935.XA
Other languages
Chinese (zh)
Other versions
CN113919506A (en
Inventor
王阳阳
吴良庆
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Jingdong Technology Information Technology Co Ltd
Original Assignee
Jingdong Technology Information Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Jingdong Technology Information Technology Co Ltd filed Critical Jingdong Technology Information Technology Co Ltd
Priority to CN202111175935.XA priority Critical patent/CN113919506B/en
Publication of CN113919506A publication Critical patent/CN113919506A/en
Application granted granted Critical
Publication of CN113919506B publication Critical patent/CN113919506B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/36Prevention of errors by analysis, debugging or testing of software
    • G06F11/3668Testing of software
    • G06F11/3672Test management
    • G06F11/3688Test management for test execution, e.g. scheduling of test suites
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5005Allocation of resources, e.g. of the central processing unit [CPU] to service a request
    • G06F9/5011Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resources being hardware resources other than CPUs, Servers and Terminals
    • G06F9/5016Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resources being hardware resources other than CPUs, Servers and Terminals the resource being the memory
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5005Allocation of resources, e.g. of the central processing unit [CPU] to service a request
    • G06F9/5011Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resources being hardware resources other than CPUs, Servers and Terminals
    • G06F9/5022Mechanisms to release resources

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Software Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • Quality & Reliability (AREA)
  • Computer Hardware Design (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Medical Informatics (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The application discloses a model training online method, a device, a storage medium and equipment, under the condition of receiving a test version number input by a user, and identifying the model to be online corresponding to the test version number as a test model, and identifying the evaluation value of the test model as a test evaluation value. And under the condition that the test evaluation value is larger than the target evaluation value and a test instruction of a user is received, starting a preset algorithm service to perform AB test on the test model and the current online model to obtain the test result of the test model and the test result of the current online model. And under the condition that the test result of the test model is higher than that of the current online model, releasing the memory of the current online model, and updating the state of the test model to be online. By utilizing the scheme disclosed by the application, an information synchronization mechanism is not required to be adopted in the training process and the online process of the model, and manual participation is not required, so that the labor cost of online training of the model can be effectively reduced.

Description

Model training online method, device, storage medium and equipment
Technical Field
The application relates to the field of intelligent customer service dialogue systems, in particular to a model training online method, device, storage medium and equipment.
Background
In intelligent customer service dialogue systems, a variety of models are generally used, and more common models include text classification, semantic matching, entity recognition and other types. With the development of natural language processing (Neuro-Linguistic Programming, NLP) algorithms, the above several types of models gradually evolve from initial statistical learning into deep learning algorithms. The performance requirements on the machine are higher and higher when the models are trained, and the common central processing unit (Central Processing Unit, CPU) machine cannot meet the deep learning requirement, so that many large companies start to purchase the machines of the GPU and the TPU for training the deep learning model.
If the production environment is to employ a large-structure model, graphics processor (Graphics Processing Unit, GPU), tensor processor (TensorProcessing Unit, TPU) resources are required for training. Generally, the machine of the GPU is mainly an offline machine, and because many enterprises isolate the production environment, the machine of the production environment cannot obtain data of offline service, and the offline service cannot obtain data of the production environment. Therefore, this situation requires a method of interaction between the production environment and the offline GPU machine to achieve model training on-line. For this approach, it is common practice to isolate the model training from the online, debug the model manually and continuously to obtain the final model, and then subject the final model and the model currently in use to AB testing to determine whether to online the final model.
Obviously, the training process and the online process of the final model need to adopt an information synchronization mechanism, however, most of the information synchronization mechanisms are manually completed by manpower, the complexity of a production environment can be improved by adopting the information synchronization mechanism, a large amount of computing resources are consumed in the AB test process, and the online process needs to be additionally developed for different types of models and businesses, so that the labor cost is greatly increased.
Disclosure of Invention
The application provides a model training online method, a device, a storage medium and equipment, and aims to reduce labor cost of model training online.
In order to achieve the above object, the present application provides the following technical solutions:
a model training online method, comprising:
Under the condition that a test version number input by a user is received, marking a model to be online corresponding to the test version number as a test model, and marking an evaluation value of the test model as a test evaluation value, wherein the model to be online and the evaluation value of the model to be online are obtained by training a model to be trained;
Under the condition that the test evaluation value is larger than a target evaluation value and a test instruction of a user is received, starting a preset algorithm service to conduct AB test on the test model and the current online model to obtain a test result of the test model and a test result of the current online model;
And under the condition that the test result of the test model is higher than the test result of the current online model, releasing the memory of the current online model, and updating the state of the test model to be online.
Optionally, the model to be online and the evaluation value of the model to be online are all obtained by training the model to be trained, and the method includes:
creating a training task of the model to be trained based on the type and the parameters of the model to be trained, and storing the training corpus of the model to be trained to a cloud;
The training task is sent to an offline graphic processor, the offline graphic processor is triggered to acquire the training corpus from the cloud end, and the model to be trained is trained according to the type to be trained, the parameters and the training corpus to obtain the model to be online and an evaluation value of the model to be online;
Setting a state label and a version number for the to-be-online model, and storing information of the to-be-online model into a preset model table.
Optionally, the creating the training task of the model to be trained based on the type and the parameters of the model to be trained includes:
Obtaining the type, parameters and training corpus of a model to be trained;
Setting a state label for the model to be trained, and storing training corpus of the model to be trained to a cloud;
Storing the type, the parameters and the state labels of the model to be trained into a database;
under the condition that an access request sent by an offline graphic processor is received, a training task of the model to be trained is created based on the type and the parameters of the model to be trained, and the state of the model to be trained is updated to be in training.
Optionally, setting a status tag and a version number for the to-be-online model, and storing information of the to-be-online model into a preset model table, including:
setting a version number for the to-be-online model under the condition that the to-be-online model and the evaluation value sent by the offline graphics processor are received;
storing the model to be online and the evaluation value into a database, and updating the state of the model to be trained to be training completion;
setting a state label for the to-be-online model, and storing information of the to-be-online model into a preset model table;
and storing the model to be online to a cloud end, and displaying the model to be online and the evaluation value to a user through a preset front-end interface.
Optionally, under the condition that the test version number input by the user is received, identifying the to-be-online model corresponding to the test version number as a test model, and identifying the evaluation value of the test model as a test evaluation value, including:
the offline graphics processor sends the offline models to be online and the evaluation values of the offline models to be online in advance to a database;
under the condition that a test version number input by a user is received, selecting a to-be-online model corresponding to the test version number from the database, marking the to-be-online model as a test model, and marking an evaluation value of the test model as a test evaluation value;
And updating the state of the test model to be tested.
Optionally, under the condition that the test evaluation value is greater than the target evaluation value and a test instruction of a user is received, starting a preset algorithm service to perform an AB test on the test model and the current online model to obtain a test result of the test model and a test result of the current online model, including:
comparing the test evaluation value with a target evaluation value;
Judging whether the current online model is the test model or not under the condition that the test evaluation value is larger than the target evaluation value;
If the current online model is the test model, prompting a user that the test model is in an online state, and updating the state of the test model to be online;
If the current online model is not the test model, under the condition that a test instruction of the user is received, starting a preset algorithm service to conduct AB test on the test model and the current online model, and obtaining test results of the test model and the current online model.
Optionally, the method further comprises:
And under the condition that the test performance of the test model is lower than that of the current online model, releasing the internal memory of the test model, and deleting the information of the test model from the model table.
A model training on-line device, comprising:
The system comprises an identification unit, a test model, a test evaluation unit and a training unit, wherein the identification unit is used for identifying a model to be online corresponding to a test version number as a test model and identifying the evaluation value of the test model as a test evaluation value under the condition that the test version number input by a user is received;
The test unit is used for starting a preset algorithm service to conduct AB test on the test model and the current online model under the condition that the test evaluation value is larger than a target evaluation value and a test instruction of a user is received, so as to obtain a test result of the test model and a test result of the current online model;
And the online unit is used for releasing the memory of the current online model and updating the state of the test model into online under the condition that the test result of the test model is higher than the test result of the current online model.
A computer readable storage medium comprising a stored program, wherein the program performs the model training on-line method.
The model training online equipment comprises a processor, a memory and a bus, wherein the processor is connected with the memory through the bus;
The memory is used for storing a program, and the processor is used for running the program, wherein the model training online method is executed when the program runs.
According to the technical scheme provided by the application, under the condition that the test version number input by the user is received, the to-be-online model corresponding to the test version number is identified as a test model, and the evaluation value of the test model is identified as a test evaluation value. And under the condition that the test evaluation value is larger than the target evaluation value and a test instruction of a user is received, starting a preset algorithm service to perform AB test on the test model and the current online model to obtain the test result of the test model and the test result of the current online model. The target evaluation value is obtained based on training a current online model by using a test set. And under the condition that the test result of the test model is higher than that of the current online model, releasing the memory of the current online model, and updating the state of the test model to be online. By utilizing the scheme disclosed by the application, an information synchronization mechanism is not required to be adopted in the training process and the online process of the model, and manual participation is not required, so that the labor cost of online training of the model is effectively reduced.
Drawings
In order to more clearly illustrate the embodiments of the application or the technical solutions in the prior art, the drawings that are required in the embodiments or the description of the prior art will be briefly described, it being obvious that the drawings in the following description are only some embodiments of the application, and that other drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
FIG. 1 is a schematic diagram of a model training online method according to an embodiment of the present application;
FIG. 2 is a schematic diagram of another model training online method according to an embodiment of the present application;
Fig. 3 is a schematic diagram of a model training online device according to an embodiment of the present application.
Detailed Description
The following description of the embodiments of the present application will be made clearly and completely with reference to the accompanying drawings, in which it is apparent that the embodiments described are only some embodiments of the present application, but not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the application without making any inventive effort, are intended to be within the scope of the application.
As shown in fig. 1, a schematic diagram of a model training online method according to an embodiment of the present application is applied to a production environment, and includes the following steps:
s101, obtaining the type, parameters and training corpus of the model to be trained.
Parameters include, but are not limited to, training algorithms, training tasks, maximum length of training corpus, whether to splice the above information, etc.
S102, setting a state label for the model to be trained, and storing training corpus of the model to be trained to a cloud end.
The training process of the model to be trained is performed in an offline environment (i.e., an offline graphics processor), that is, the offline graphics processor cannot directly access the database. For this reason, the training corpus of the model to be trained needs to be stored in the cloud end, so as to be convenient for the offline graphics processor to acquire.
And S103, storing the type, the parameters and the state labels of the model to be trained into a database.
And S104, under the condition that an access request sent by the offline graphic processor is received, creating a training task of the model to be trained based on the type and the parameters of the model to be trained, and updating the state of the model to be trained into training.
The offline graphics processor can send an access request through a preset http interface. In addition, by pre-deploying a timing access service on the offline graphics processor, the offline graphics processor can be controlled to send access requests at a timing.
And S105, transmitting the training task to the offline graphic processor, triggering the offline graphic processor to acquire training corpus of the model to be trained from the cloud, and training the model to be trained according to the type, the parameters and the training corpus of the model to be trained to obtain a model file and a result file.
The model file comprises a model to be online, the model to be trained is obtained through training of the model to be trained, and the result file comprises an evaluation value for evaluating the training effect of the model to be online, wherein the evaluation value is obtained through training of the model to be online by using a test set.
It should be noted that, the offline graphics processor will generally pre-configure a resource table, where the resource table is used to record the usage states of its own computing resources, and before executing the training task, the offline graphics processor will query the usage states of the computing resources in the resource table, and if there are computing resources whose usage states are idle, identify the computing resources whose usage states are idle as target resources, and execute the training task by using the target resources.
Generally, the offline graphic processor can realize the training process of the model to be trained by calling a pre-deployed training process, specifically, in codes shown in the training process, the type, parameters and training corpus of the model to be trained are configured, and the training process is operated to generate a model file and a result file.
In addition, the offline graphics processing may also identify the use status of the target resource as in-use during the execution of the training task using the target resource. After the model file and the result file are obtained, i.e., after the training task is finished, the offline graphics processor also identifies the use state of the target resource as idle.
And S106, setting a version number for the to-be-online model under the condition that the to-be-online model and the evaluation value sent by the offline graphics processor are received.
And S107, storing the model to be online and the evaluation value into a database, and updating the state of the model to be trained to be training completion.
S108, setting a state label for the to-be-online model, and storing information of the to-be-online model into a preset model table.
And S109, storing the model to be online to a cloud end, and displaying the model to be online and the evaluation value to a user through a preset front end interface.
The process shown in S101-S109 mainly describes that the production environment issues a training task of the model to be trained to the offline graphics processor, and after the offline graphics processor completes the training task, the model file and the result file obtained by training are fed back to the production environment, so as to realize automatic training of the model to be trained. Compared with the prior art, the training process of the model to be trained does not need to be manually participated, and the labor cost of model training on line is effectively reduced.
S110, under the condition that the test version number input by the user is received, selecting a model to be online corresponding to the test version number from the database, and marking the model as a test model.
And S111, identifying the evaluation value of the test model as a test evaluation value.
S112, updating the state of the test model to be tested.
And S113, comparing the test evaluation value with the target evaluation value.
The target evaluation value is obtained based on training a current online model by using a test set. In the embodiment of the application, the target evaluation value is used for evaluating the training effect of the current online model.
And S114, judging whether the current online model is a test model or not under the condition that the test evaluation value is larger than the target evaluation value.
If the current online model is a test model, S115 is executed, otherwise S116 is executed.
In order to avoid redundant database operations caused by repeated online of the test model, whether the current online model is the test model needs to be judged, if the current online model is the test model, the test model is in an online state, and online operations are not needed.
S115, prompting the user that the test model is in the online state, and updating the state of the test model to be online.
And S116, under the condition that a test instruction of a user is received, starting a preset algorithm service to perform AB test on the test model and the current online model, so as to obtain the test result of the test model and the test result of the current online model.
After S116 is performed, S117 is continued.
Wherein the algorithm service is configured to perform the following logic:
1. And (5) regularly inquiring the model table, and judging whether the model table contains the information of the test model.
2. Under the condition that the model table contains the information of the test model, loading the test model obtained from the cloud, adding first buried point information into the current online model, and adding second buried point information into the test model.
3. And splitting the flow of the current service according to the message id to obtain a first flow and a second flow, enabling the first flow to call the current online model, and enabling the second flow to call the test model.
4. And (3) associating information shown in different message ids through the first embedded point information to obtain a service index of the first flow, and taking the value of the service index of the first flow as the test result of the current online model.
5. And (3) correlating the information shown by different message ids through the second embedded point information to obtain service indexes of the second flow, and taking the values of the service indexes of the second flow as test results of the test model.
6. And sending a prompt of AB test failure to a user when the model table does not contain information of the test model.
S117, comparing the test result of the test model with the test result of the current online model.
And S118, under the condition that the test result of the test model is higher than that of the current online model, releasing the memory of the current online model, and updating the state of the test model to be online.
And S119, under the condition that the test result of the test model is lower than the test result of the current online model, releasing the internal memory of the test model, and deleting the information of the test model from the model table.
And S120, under the condition that the test model is detected to be a new model, loading the test model, and updating the state of the test model to be on line.
In summary, by using the scheme shown in the embodiment, the training process and the online process of the model do not need to adopt an information synchronization mechanism, and do not need to be manually participated, so that the labor cost of model training and online is effectively reduced.
It should be noted that S119 mentioned in the foregoing embodiment is an alternative implementation of the model training online method shown in the present application. In addition, S120 mentioned in the foregoing embodiment is also an alternative implementation of the model training online method shown in the present application. For this reason, the flow mentioned in the above embodiment can be summarized as the method shown in fig. 2.
As shown in fig. 2, a schematic diagram of another model training online method according to an embodiment of the present application includes the following steps:
And S201, under the condition that the test version number input by the user is received, identifying the model to be online corresponding to the test version number as a test model, and identifying the evaluation value of the test model as a test evaluation value.
The model to be online and the evaluation value of the model to be online are obtained by training the model to be trained.
And S202, under the condition that the test evaluation value is larger than the target evaluation value and a test instruction of a user is received, starting a preset algorithm service to perform AB test on the test model and the current online model to obtain the test result of the test model and the test result of the current online model.
The target evaluation value is obtained based on training a current online model by using a test set.
And S203, under the condition that the test result of the test model is higher than that of the current online model, releasing the memory of the current online model, and updating the state of the test model to be online.
In summary, by using the scheme shown in the embodiment, the training process and the online process of the model do not need to adopt an information synchronization mechanism, and do not need to be manually participated, so that the labor cost of model training and online is effectively reduced.
Corresponding to the model training online method provided by the embodiment of the application, the embodiment of the application also provides a model training online device.
As shown in fig. 3, an architecture diagram of a model training online device according to an embodiment of the present application includes:
the identifying unit 100 is configured to identify, when a test version number input by a user is received, a model to be online corresponding to the test version number as a test model, and an evaluation value of the test model as a test evaluation value.
The identification unit 100 is specifically configured to create a training task of a model to be trained based on a type and parameters of the model to be trained, store training corpus of the model to be trained to a cloud, send the training task to an offline graphics processor, trigger the offline graphics processor to acquire the training corpus from the cloud, train the model to be trained according to the type, parameters and training corpus to obtain an online model to be trained, and an evaluation value of the online model to be trained, wherein the evaluation value is obtained based on training the online model by using a test set, set a state tag and a version number for the online model to be trained, and store information of the online model to a preset model table.
The identification unit 100 is specifically configured to obtain a type, a parameter and a training corpus of a model to be trained, set a state label for the model to be trained, store the training corpus of the model to be trained in a cloud, store the type, the parameter and the state label of the model to be trained in a database, create a training task of the model to be trained based on the type and the parameter of the model to be trained when an access request sent by an offline graphics processor is received, and update the state of the model to be trained in training.
The identification unit 100 is specifically configured to set a version number for the to-be-online model under the condition that the to-be-online model and the evaluation value sent by the offline graphics processor are received, store the to-be-online model and the evaluation value in a database, update the state of the to-be-trained model to be training completion, set a state label for the to-be-online model, store information of the to-be-online model in a preset model table, store the to-be-online model to a cloud, and display the to-be-online model and the evaluation value to a user through a preset front end interface.
The identification unit 100 is specifically configured to store, in advance, an offline model and an evaluation value of the offline model sent by the offline graphics processor into a database, select, from the database, the offline model corresponding to the test version number, identify the online model as a test model, identify the evaluation value of the test model as a test evaluation value, and update a state of the test model as a to-be-tested under the condition that the test version number input by a user is received.
And the test unit 200 is used for starting a preset algorithm service to perform AB test on the test model and the current online model under the condition that the test evaluation value is larger than the target evaluation value and a test instruction of a user is received, so as to obtain the test result of the test model and the test result of the current online model, and the target evaluation value is obtained by training the current online model by using the test set.
The test unit 200 is specifically configured to compare the test evaluation value with the target evaluation value, determine whether the current online model is a test model if the test evaluation value is greater than the target evaluation value, prompt the user that the test model is in an online state and update the state of the test model to be online if the current online model is the test model, and start a preset algorithm service to perform an AB test on the test model and the current online model to obtain a test result of the test model and a test result of the current online model if the current online model is not the test model, under the condition that a test instruction of the user is received.
And the online unit 300 is configured to perform memory release on the current online model and update the state of the test model to online when the test result of the test model is higher than the test result of the current online model.
And the deleting unit 400 is configured to release the memory of the test model and delete the information of the test model from the model table when the test result of the test model is lower than the test result of the current online model.
In summary, by using the scheme shown in the embodiment, the training process and the online process of the model do not need to adopt an information synchronization mechanism, and do not need to be manually participated, so that the labor cost of model training and online is effectively reduced.
The application also provides a computer readable storage medium, wherein the computer readable storage medium comprises a stored program, and the program executes the model training online method provided by the application.
The application also provides model training online equipment which comprises a processor, a memory and a bus. The processor is connected with the memory through a bus, the memory is used for storing a program, and the processor is used for running the program, wherein the model training online method provided by the application is executed when the program runs, and the method comprises the following steps:
Under the condition that a test version number input by a user is received, marking a model to be online corresponding to the test version number as a test model, and marking an evaluation value of the test model as a test evaluation value, wherein the model to be online and the evaluation value of the model to be online are obtained by training a model to be trained;
Under the condition that the test evaluation value is larger than a target evaluation value and a test instruction of a user is received, starting a preset algorithm service to conduct AB test on the test model and the current online model to obtain a test result of the test model and a test result of the current online model;
And under the condition that the test result of the test model is higher than the test result of the current online model, releasing the memory of the current online model, and updating the state of the test model to be online.
Optionally, the model to be online and the evaluation value of the model to be online are all obtained by training the model to be trained, and the method includes:
creating a training task of the model to be trained based on the type and the parameters of the model to be trained, and storing the training corpus of the model to be trained to a cloud;
The training task is sent to an offline graphic processor, the offline graphic processor is triggered to acquire the training corpus from the cloud end, and the model to be trained is trained according to the type to be trained, the parameters and the training corpus to obtain the model to be online and an evaluation value of the model to be online;
Setting a state label and a version number for the to-be-online model, and storing information of the to-be-online model into a preset model table.
Optionally, the creating the training task of the model to be trained based on the type and the parameters of the model to be trained includes:
Obtaining the type, parameters and training corpus of a model to be trained;
Setting a state label for the model to be trained, and storing training corpus of the model to be trained to a cloud;
Storing the type, the parameters and the state labels of the model to be trained into a database;
under the condition that an access request sent by an offline graphic processor is received, a training task of the model to be trained is created based on the type and the parameters of the model to be trained, and the state of the model to be trained is updated to be in training.
Optionally, setting a status tag and a version number for the to-be-online model, and storing information of the to-be-online model into a preset model table, including:
setting a version number for the to-be-online model under the condition that the to-be-online model and the evaluation value sent by the offline graphics processor are received;
storing the model to be online and the evaluation value into a database, and updating the state of the model to be trained to be training completion;
setting a state label for the to-be-online model, and storing information of the to-be-online model into a preset model table;
and storing the model to be online to a cloud end, and displaying the model to be online and the evaluation value to a user through a preset front-end interface.
Optionally, under the condition that the test version number input by the user is received, identifying the to-be-online model corresponding to the test version number as a test model, and identifying the evaluation value of the test model as a test evaluation value, including:
the offline graphics processor sends the offline models to be online and the evaluation values of the offline models to be online in advance to a database;
under the condition that a test version number input by a user is received, selecting a to-be-online model corresponding to the test version number from the database, marking the to-be-online model as a test model, and marking an evaluation value of the test model as a test evaluation value;
And updating the state of the test model to be tested.
Optionally, under the condition that the test evaluation value is greater than the target evaluation value and a test instruction of a user is received, starting a preset algorithm service to perform an AB test on the test model and the current online model to obtain a test result of the test model and a test result of the current online model, including:
comparing the test evaluation value with a target evaluation value;
Judging whether the current online model is the test model or not under the condition that the test evaluation value is larger than the target evaluation value;
If the current online model is the test model, prompting a user that the test model is in an online state, and updating the state of the test model to be online;
If the current online model is not the test model, under the condition that a test instruction of the user is received, starting a preset algorithm service to conduct AB test on the test model and the current online model, and obtaining test results of the test model and the current online model.
Optionally, the method further comprises:
And under the condition that the test performance of the test model is lower than that of the current online model, releasing the internal memory of the test model, and deleting the information of the test model from the model table.
The functions of the methods of embodiments of the present application, if implemented in the form of software functional units and sold or used as a stand-alone product, may be stored on a computing device readable storage medium. Based on such understanding, a part of the present application that contributes to the prior art or a part of the technical solution may be embodied in the form of a software product stored in a storage medium, comprising several instructions for causing a computing device (which may be a personal computer, a server, a mobile computing device or a network device, etc.) to execute all or part of the steps of the method described in the embodiments of the present application. The storage medium includes a U disk, a removable hard disk, a Read-Only Memory (ROM), a random access Memory (RAM, random Access Memory), a magnetic disk, an optical disk, or other various media capable of storing program codes.
In this specification, each embodiment is described in a progressive manner, and each embodiment is mainly described in a different point from other embodiments, so that the same or similar parts between the embodiments are referred to each other.
The previous description of the disclosed embodiments is provided to enable any person skilled in the art to make or use the present application. Various modifications to these embodiments will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other embodiments without departing from the spirit or scope of the application. Thus, the present application is not intended to be limited to the embodiments shown herein but is to be accorded the widest scope consistent with the principles and novel features disclosed herein.

Claims (10)

1. A model training on-line method, comprising:
Under the condition that a test version number input by a user is received, marking a model to be online corresponding to the test version number as a test model, and marking an evaluation value of the test model as a test evaluation value, wherein the model to be online and the evaluation value of the model to be online are obtained by training a model to be trained;
Under the condition that the test evaluation value is larger than a target evaluation value and a test instruction of a user is received, starting a preset algorithm service to conduct AB test on the test model and the current online model to obtain a test result of the test model and a test result of the current online model;
And under the condition that the test result of the test model is higher than the test result of the current online model, releasing the memory of the current online model, and updating the state of the test model to be online.
2. The method of claim 1, wherein the model to be online and the evaluation value of the model to be online are obtained by training a model to be trained, comprising:
creating a training task of the model to be trained based on the type and the parameters of the model to be trained, and storing the training corpus of the model to be trained to a cloud;
The training task is sent to an offline graphic processor, the offline graphic processor is triggered to acquire the training corpus from the cloud end, and the model to be trained is trained according to the type to be trained, the parameters and the training corpus to obtain the model to be online and an evaluation value of the model to be online;
Setting a state label and a version number for the to-be-online model, and storing information of the to-be-online model into a preset model table.
3. The method of claim 2, wherein creating the training task for the model to be trained based on the type and parameters of the model to be trained comprises:
Obtaining the type, parameters and training corpus of a model to be trained;
Setting a state label for the model to be trained, and storing training corpus of the model to be trained to a cloud;
Storing the type, the parameters and the state labels of the model to be trained into a database;
under the condition that an access request sent by an offline graphic processor is received, a training task of the model to be trained is created based on the type and the parameters of the model to be trained, and the state of the model to be trained is updated to be in training.
4. The method of claim 2, wherein the setting a status tag and a version number for the to-be-online model and storing information of the to-be-online model in a preset model table includes:
setting a version number for the to-be-online model under the condition that the to-be-online model and the evaluation value sent by the offline graphics processor are received;
storing the model to be online and the evaluation value into a database, and updating the state of the model to be trained to be training completion;
setting a state label for the to-be-online model, and storing information of the to-be-online model into a preset model table;
and storing the model to be online to a cloud end, and displaying the model to be online and the evaluation value to a user through a preset front-end interface.
5. The method according to claim 2, wherein in the case of receiving a test version number input by a user, identifying a model to be online corresponding to the test version number as a test model, and identifying an evaluation value of the test model as a test evaluation value, includes:
the offline graphics processor sends the offline models to be online and the evaluation values of the offline models to be online in advance to a database;
under the condition that a test version number input by a user is received, selecting a to-be-online model corresponding to the test version number from the database, marking the to-be-online model as a test model, and marking an evaluation value of the test model as a test evaluation value;
And updating the state of the test model to be tested.
6. The method according to claim 1, wherein, when the test evaluation value is greater than the target evaluation value and a test instruction of a user is received, starting a preset algorithm service to perform an AB test on the test model and the current online model, to obtain a test result of the test model and a test result of the current online model, including:
comparing the test evaluation value with a target evaluation value;
Judging whether the current online model is the test model or not under the condition that the test evaluation value is larger than the target evaluation value;
If the current online model is the test model, prompting a user that the test model is in an online state, and updating the state of the test model to be online;
If the current online model is not the test model, under the condition that a test instruction of the user is received, starting a preset algorithm service to conduct AB test on the test model and the current online model, and obtaining test results of the test model and the current online model.
7. The method as recited in claim 1, further comprising:
And under the condition that the test performance of the test model is lower than that of the current online model, releasing the internal memory of the test model, and deleting the information of the test model from the model table.
8. A model training on-line device, comprising:
The system comprises an identification unit, a test model, a test evaluation unit and a training unit, wherein the identification unit is used for identifying a model to be online corresponding to a test version number as a test model and identifying the evaluation value of the test model as a test evaluation value under the condition that the test version number input by a user is received;
The test unit is used for starting a preset algorithm service to conduct AB test on the test model and the current online model under the condition that the test evaluation value is larger than a target evaluation value and a test instruction of a user is received, so as to obtain a test result of the test model and a test result of the current online model;
And the online unit is used for releasing the memory of the current online model and updating the state of the test model into online under the condition that the test result of the test model is higher than the test result of the current online model.
9. A computer readable storage medium, characterized in that the computer readable storage medium comprises a stored program, wherein the program performs the model training on-line method of any of claims 1-7.
10. The model training online equipment is characterized by comprising a processor, a memory and a bus, wherein the processor is connected with the memory through the bus;
The memory is used for storing a program, and the processor is used for running the program, wherein the program executes the model training online method according to any one of claims 1-7.
CN202111175935.XA 2021-10-09 2021-10-09 A model training online method, device, storage medium and equipment Active CN113919506B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111175935.XA CN113919506B (en) 2021-10-09 2021-10-09 A model training online method, device, storage medium and equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111175935.XA CN113919506B (en) 2021-10-09 2021-10-09 A model training online method, device, storage medium and equipment

Publications (2)

Publication Number Publication Date
CN113919506A CN113919506A (en) 2022-01-11
CN113919506B true CN113919506B (en) 2024-12-24

Family

ID=79238720

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111175935.XA Active CN113919506B (en) 2021-10-09 2021-10-09 A model training online method, device, storage medium and equipment

Country Status (1)

Country Link
CN (1) CN113919506B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113722394A (en) * 2021-08-17 2021-11-30 北京百悟科技有限公司 Data synchronization method, device and storage medium
CN116383026B (en) * 2023-06-05 2023-09-01 阿里巴巴(中国)有限公司 Data processing method and server based on large model

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112288092A (en) * 2019-07-23 2021-01-29 百度时代网络技术(北京)有限公司 Model evaluation method, model evaluation device, electronic device and storage medium
CN113065843A (en) * 2021-03-15 2021-07-02 腾讯科技(深圳)有限公司 Model processing method and device, electronic equipment and storage medium

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11176632B2 (en) * 2017-04-07 2021-11-16 Intel Corporation Advanced artificial intelligence agent for modeling physical interactions
CN111753843A (en) * 2020-06-28 2020-10-09 平安科技(深圳)有限公司 Segmentation effect evaluation method, device, equipment and medium based on deep learning

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112288092A (en) * 2019-07-23 2021-01-29 百度时代网络技术(北京)有限公司 Model evaluation method, model evaluation device, electronic device and storage medium
CN113065843A (en) * 2021-03-15 2021-07-02 腾讯科技(深圳)有限公司 Model processing method and device, electronic equipment and storage medium

Also Published As

Publication number Publication date
CN113919506A (en) 2022-01-11

Similar Documents

Publication Publication Date Title
CN113919506B (en) A model training online method, device, storage medium and equipment
US20220052976A1 (en) Answer text processing methods and apparatuses, and key text determination methods
CN110991871A (en) Risk monitoring method, device, equipment and computer readable storage medium
CN113742226B (en) Software performance test method and device, medium and electronic equipment
CN112784273A (en) A method, device and equipment for SQL risk identification
CN113051178A (en) Test data construction method, device, equipment and medium
CN112100239A (en) Portrait generation method and apparatus for vehicle detection device, server and readable storage medium
CN115759372A (en) Process engine optimization method and system based on business processing and process driving decoupling
CN110990285A (en) UI automation test method and device
CN110941486A (en) Task management method and device, electronic equipment and computer readable storage medium
CN103701671B (en) Method and device for detecting conflicts among businesses
CN110674275B (en) Knowledge question answering method and device
CN110874319A (en) Automated testing method, automated testing platform, automated testing equipment and computer-readable storage medium
CN118503396A (en) ERP system large model calling method, device and medium based on open prompt words
CN110209565A (en) A kind of metadata schema adjustment method and its device
CN117370356A (en) Method and related device for mapping metadata by data standard
CN112579833B (en) Service association relation acquisition method and device based on user operation data
CN113672497B (en) Method, device and equipment for generating non-buried point event and storage medium
CN115757186A (en) Performance test method, device, equipment and medium of software interface
CN111562982B (en) Method and device for processing request data, computer readable storage medium and electronic equipment
CN114186046A (en) Information processing method, information processing apparatus, server, and storage medium
CN115273858A (en) A text processing method, device, equipment and medium
CN114417150A (en) Recommended methods, devices, equipment and storage media for troubleshooting steps
CN111008140A (en) A cross-platform UI automation testing method and device
CN114692647B (en) Data processing method, device, equipment and medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant