US11474689B2 - Method and apparatus for obtaining content - Google Patents
Method and apparatus for obtaining content Download PDFInfo
- Publication number
- US11474689B2 US11474689B2 US17/136,667 US202017136667A US11474689B2 US 11474689 B2 US11474689 B2 US 11474689B2 US 202017136667 A US202017136667 A US 202017136667A US 11474689 B2 US11474689 B2 US 11474689B2
- Authority
- US
- United States
- Prior art keywords
- swiping
- page
- distance
- parameters
- speed
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
- G06F3/04883—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/0483—Interaction with page-structured environments, e.g. book metaphor
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/0485—Scrolling or panning
Definitions
- the disclosure relates to a field of Internet technologies, and more particularly, to a method and an apparatus for obtaining content, an electronic device and a storage medium.
- a method for obtaining content includes:
- an apparatus for obtaining content includes a processor and a memory.
- the memory is configured to store instructions executable by the processor.
- the processor is configured to execute the instructions stored in the memory, so as to:
- an electronic device includes:
- a memory configured to store instructions executable by the processor
- processor configured to execute the instructions, so as to implement the method in any embodiment of the first aspect.
- a non-transitory computer-readable storage medium When instructions in the storage medium are executed by a processor of an electronic device, the electronic device is caused to implement the method in any embodiment of the first aspect.
- FIG. 1 is a schematic diagram illustrating an application environment of a method for obtaining content according to an embodiment.
- FIG. 2 is a flowchart of a method for obtaining content according to an embodiment.
- FIG. 3 is a flowchart of determining an object finally positioned according to an embodiment.
- FIG. 4 is flowchart of determining a page swiping distance according to an embodiment.
- FIG. 5 is a flowchart of determining an initial page swiping distance according to an embodiment.
- FIG. 6 is a flowchart of adjusting a page according to an embodiment.
- FIG. 7 is a flowchart of a method for loading content according to an embodiment.
- FIG. 8 is a schematic diagram of a page according to an embodiment.
- FIG. 9 is a block diagram of an apparatus for obtaining content according to an embodiment.
- FIG. 10 is a block diagram of an apparatus for obtaining content according to another embodiment.
- FIG. 11 is a block diagram of an electronic device according to an embodiment.
- short video content is usually obtained by preloading part of the content in a coarse-grained manner and then loading detailed content of the target short video in response to receiving a content obtaining instruction from the user for the target short video. Then, the short video content is sent to the client for the user to browse.
- This kind of loading may cause the client to wait for a certain time in the process of obtaining the page content, which leads to the problem of low efficiency.
- embodiments of the disclosure provide a method for obtaining content.
- the method for obtaining content provided by the disclosure may be applied in the application environment as illustrated in FIG. 1 .
- the terminal 110 interacts with the server 120 via the network.
- the terminal 110 is installed with various applications, such as social applications and instant messaging applications.
- the terminal 110 includes a screen for human-machine interaction and for receiving operations on a page, so that the terminal may display the page and objects in the page based on the corresponding operation. For example, when the terminal receives a swiping-down operation through the screen, the terminal may move the page downwards based on the swiping-down operation.
- the terminal may display content of the clicked object based on the click operation.
- the objects include, but not limited to, videos, pictures and text published by the user.
- the server 120 stores content of respective objects of the page, and the terminal may obtain the content of respective objects from the server 120 and display it through the screen.
- the terminal obtains a swiping speed of a swiping in response to detecting an end of the swiping applied on the page, determines the object finally positioned by the swiping in the page based on the swiping speed and page parameters of the page, and obtains and displays the content of the object.
- the terminal 110 may be but not limited to a variety of personal computers, laptops, smart phones, tablets and portable wearable devices.
- the server 120 may be implemented as an independent server or a server cluster composed of multiple servers.
- FIG. 2 is a flowchart of a method for obtaining content according to an embodiment. As illustrated in FIG. 2 , the method is applied in the terminal 110 , and includes following actions.
- a swiping speed of the swiping is obtained.
- the page is displayed on a screen of the terminal 110 and a user is operating on the page.
- the page may include at least one object currently displayed in the screen and several objects out of the screen obtained by a coarse-grained preloading.
- the terminal moves the page accordingly based on the direction of the swiping, to display the objects in the page, so that the user may continuously browse the objects in the page.
- the direction of the swiping is not limited to any of upwards, downwards, towards the left, and towards the right.
- the terminal detects the end of the swiping applied on the page, the terminal obtains the swiping speed of the swiping.
- the swiping speed may be obtained by the terminal through performing swiping speed tracking in real time. That is, the terminal may obtain the swiping speed by tracking a series of events, for example, tracking the distance (for example, the number of pixels) which the swiping passes through in unit time (for example, in the time unit of 1 millisecond).
- the swiping speed may also be detected in real time by sensors installed in the terminal, which is not limited here.
- an object positioned by the swiping in the page is determined based on the swiping speed and page parameters of the page.
- the object positioned by the swiping refers to any object that can be displayed in the screen when the page stops moving.
- the object positioned by the swiping is the last object displayed in the screen along the swiping direction. For example, when the swiping is upward, and the page moves upwards therewith, then the object positioned by the swiping may refer to the object displayed at the lowest position of the screen when the page stops moving.
- the page parameters may refer to various parameters included in the page, for example, display parameters of the page, addresses of respective objects in the page, and screen positions of the respective objects.
- the terminal determines the object displayed in the screen when the page stops moving, based on the swiping speed and the page parameters. Further, all the content of the object positioned by the swiping is loaded, such that the user may directly read all the content of the object positioned by the swiping when the page stops moving.
- the object to be displayed in the screen when the page is static is positioned precisely based on the obtained swiping speed of the swiping and page parameters of the page.
- Detailed content of the object is obtained in advance, such that the user can directly browse the detailed content of the object when the page is static, without additional waiting, thereby improving the efficiency of obtaining content of the object.
- by loading the detailed content of the object positioned by the swiping instead of loading content of all the objects in the page, it is possible to avoid unnecessary resource waste.
- the object positioned by the swiping in the page may be determined based on the swiping speed and the page parameters of the page as follows.
- a page swiping distance is determined based on the swiping speed and the page parameters.
- the object positioned by the swiping in the page is determined based on the page swiping distance and the page parameters.
- the page swiping distance may refer to the distance generated from triggering the swiping to stopping moving the page.
- Factors affecting the page swiping distance include but not limited to, the force, the swiping speed and the number of pressings of the swiping applied on the screen.
- the page swiping distance may be determined based on the number of pressings within a specified area.
- the terminal in order to position the object in the page precisely, the terminal may determine the page swiping distance based on the swiping speed obtained at the end of the swiping and the page parameters of the page.
- the page moves accordingly.
- the terminal may determine the distance of the page moving during the swiping based on the moving distance of the object of the page in the screen.
- the terminal may determine the distance of the page moving during the swiping based on monitored parameters such as the swiping speed and the swiping time.
- the page may continue to swipe a certain distance due to inertial motion. Therefore, in view of the principle of inertial motion, an inertial swiping distance after the end of the swiping may be determined based on the swiping speed or the inertial page wiping distance obtained from multiple experiments.
- the terminal obtains the page swiping distance based on the total moving distance, and thereafter, determines the object positioned by the swiping based on the page swiping distance and the page parameters.
- the page parameters include display parameters of a plurality of objects in the page.
- the display parameters are obtained at the first loading of the plurality of objects.
- determining the object positioned by the swiping in the page based on the page swiping distance and the page parameters of the page includes: determining page display sizes of respective objects based on the display parameters of respective objects; and determining the object positioned by the swiping in the page based on the page swiping distance and the page display sizes of the respective objects.
- the display parameters refer to related parameters supporting configuring display effects of respective objects in the page.
- the display parameters are not limited to size information of respective objects, such as aspect ratios; cover content; coordinate information and preset arrangement rules of respective objects in the page.
- the cover content may be a thumbnail of the picture; when the object is a short video, the cover content may be a screenshot of the short video.
- the display parameters are obtained at first coarse-grained loading of respective objects in the page. Obtaining the display parameters at the first loading enables the respective objects in the page to be displayed in the screen, such that the user can browse the brief content of respective objects.
- the respective objects in the page may be pre-arranged based on requirements and design layout of the product, to obtain page display sizes of respective objects.
- the display sizes include but not limited to display heights and display widths.
- the display sizes may be obtained based on the requirements and design layout of the product in conjunction with aspect ratios of respective objects, or may be preset fixed sizes.
- the cover of the picture may be set to have a fixed size. For example, when the picture is shot in a landscape mode, the display heigh of the cover is H 1 , and when the picture is shot in a portrait mode, the display height of the cover is H 2 .
- the display height may change with the actual size of the picture, which is not further described here.
- the object positioned when the page stops moving may be positioned precisely based on the obtained page swiping distance and the display sizes of respective objects.
- the accuracy of positioning the object may be improved.
- the page parameters include screen position information of an initial object of the swiping in the screen.
- the initial object is in the page.
- determining the page swiping distance based on the swiping speed and the page parameters may include followings.
- an initial page swiping distance of the page during the swiping is determined based on the screen position information of the initial object.
- the initial object of the swiping refers to the object where the swiping is initially located, i.e., the object where the swiping is located at the starting of the swiping.
- the swiping is triggered by a finger of the user, and then the initial object of the swiping is the object where the finger is located.
- the terminal obtains the initial page swiping distance by monitoring change of position of the initial object in the screen.
- an inertial swiping distance of the page after the end of the swiping is determined based on the swiping speed.
- the inertial swiping distance refers to that, after the swiping, due to the inertial motion, the page may continue to swipe a certain distance within a period of time, i.e., slow down to a static state.
- the terminal may determine the distance which the page still needs to move based on the detected swiping speed, and take this distance as the inertial swiping distance.
- the inertial swiping distance may be calculated based on the swiping speed, or may be determined through multiple experiments.
- a correspondence between the swiping speeds and the inertial swiping distances is determined through multiple experiments, and after the swiping speed is obtained, the inertial swiping distance may be obtained from the correspondence. After the initial page swiping distance and the inertial swiping distance are obtained, the sum of the initial page swiping distance and the inertial swiping distance is taken as the page swiping distance. In embodiments, by including the distances generated in respective stages of the swiping of the page, the page swiping distance may be determined precisely, thereby improving the accuracy of positioning the object.
- the initial page swiping distance of the page during the swiping may be determined based on the screen position information of the initial object as follows.
- a first screen position of the initial object is obtained in response to a trigger of the swiping.
- a second screen position of the initial object is obtained in response to the end of the swiping.
- a swiping distance of the initial object in the page swiping direction is obtained based on the first screen position and the second screen position, and taken as the initial page swiping distance.
- the first screen position and the second screen position are not limited to the pixel position of the initial object in the screen.
- the terminal detects that the swiping is triggered, the terminal obtains the position of the initial object in the screen as the first screen position.
- the terminal detects that the swiping ends, the terminal obtains the position of the initial object in the screen as the second screen position.
- the swiping distance of the initial object in the page swiping direction is calculated based on the first screen position and the second screen position, and taken as the initial page swiping distance. In this embodiment, by precisely positioning the position of the initial object in the screen, the accuracy is high and it is convenient to implement.
- determining the inertial swiping distance of the page after the end of the swiping based on the swiping speed includes: obtaining the inertial swiping distance of the page after the end of the swiping based on the swiping speed and a known inertia constant.
- the inertia constant is configured in advance depending on actual requirements of the product.
- the speed of the inertial swiping of the page may be adjusted by adjusting the inertia constant.
- the inertia constant may be obtained by processing a plurality of historical inertia constants. In detail, after detecting that the swiping ends, the terminal obtains the swiping speed at the time of lifting the finger.
- the inertial swiping distance may be determined based on the following formula:
- s is the inertial swiping distance
- v is the swiping speed
- A is the inertia constant
- the inertial swiping distance is calculated based on the obtained swiping speed and the preset inertia constant, improving the accuracy.
- the swiping speed of the page may be controlled by adjusting the inertia constant, thereby having a high flexibility.
- the method further includes following actions.
- page positions of respective objects are determined based on the display parameters of the respective objects.
- a distance is obtained based on the page position of the initial object of the swiping and the page position of the object positioned by the swiping.
- the page positions refer to positions of respective objects in the page.
- the page positions may include page position coordinates in the page.
- the distance between the object where the swiping is initially located in the page and the object finally positioned by the swiping is a true distance.
- the page positions of respective objects may be determined based on the display parameters of respective objects. Then, the distance between the initial object and the object positioned by the swiping in the page is determined based on the page position of the initial object of the swiping and the page position of the object positioned by the swiping.
- the page is moved based on the offset in response to the offset being less than a threshold.
- the object positioned by the swiping may be displayed partially or fully in the screen.
- the terminal calculates the difference between the page swiping distance and the true distance as the offset.
- the offset is compared with the threshold, to determine whether the object finally positioned when the page stops moving is fully displayed in the screen. If the offset is less than the threshold, it indicates that the object cannot be fully displayed in the screen when the page stops moving. Therefore, the page may be moved continuously in the swiping direction based on the offset, thereby enabling the object to be fully displayed in the screen.
- the threshold is not limited to 0.
- the page swiping distance is compared with the distance between the initial object and the object positioned by the swiping in the page, and the page is adjusted based on the comparing result, and thus the object positioned by the swiping can be displayed fully in the screen, without the need for the user to swipe and adjust the page again, improving the convenience.
- FIG. 7 is a flowchart of a method for obtaining content according to an embodiment.
- the terminal may display the page by using the technique of feed streaming.
- the feed streaming refers to that, several message sources actively subscribed by the user are combined together to form a content aggregator, which may help the user continuously obtain the latest content of the subscribed source.
- Each object (such as state or message) in the feed streaming is one flow.
- the short videos may be updated into the feed streaming based on the chronological sequence.
- the user may continuously browse the feeds in the feed streaming.
- feed A, feed B and feed C are objects currently displayed in the screen.
- the feed streaming is moved accordingly based on the swiping, and thus the feed before feed A or the feed after feed C may be displayed.
- the feed streaming is moved upwards as an example.
- the target feed may be positioned when the feed streaming is static, with reference to the method in embodiments of the disclosure, and then the detailed content of the target feed may be obtained. As illustrated in FIG. 7 , the following actions are included.
- page parameters of the feed streaming are obtained.
- the page parameters include but not limited to, the feed ID of each feed in the feed streaming, the display parameters of each feed, and the screen position information of each feed in the screen.
- the display parameters of each feed are obtained from first preloading, including but not limited to, the aspect ratio and the display content of the cover.
- the page display sizes of respective feeds are determined based on the display parameters of respective feeds.
- the display sizes include but not limited to, the heights and widths of respective feeds. In this embodiment, take moving the feed streaming upwards as an example, the display sizes may refer to the display heights.
- the terminal after obtaining the display parameters of respective feeds, the terminal pre-arranges the respective feeds in the feed streaming based on the requirements and design of the product, thereby knowing the display heights of respective feeds.
- the first screen position information of the operation feed is obtained in response to detecting that the swiping is triggered.
- the operation feed is the feed where the swiping is located when the swiping is performed.
- the swiping is triggered by the finger of the user, and the operation feed of the swiping is the feed where the finger is located.
- the second screen position information of the operation feed is obtained and the current swiping speed of the swiping is obtained, in response to detecting the end of the swiping.
- the pixel distance which the swiping passes through in unit time may be taken as the swiping speed.
- the swiping distance of the feed streaming in the swiping direction is obtained based on the first screen position information and the second screen position information, and taken as the initial page swiping distance.
- the inertial swiping distance of the feed streaming continuously swiping after the end of the swiping is obtained based on the swiping speed and the know inertia constant.
- the inertial swiping distance may be obtained based on the following formula:
- s v * ⁇ 2 A , where, s is the inertial swiping distance, v is the swiping speed, and A is the inertia constant.
- the inertia constant may be obtained by processing a plurality of historical inertia constants.
- the sum of the initial page swiping distance and the inertial swiping distance is obtained as the page swiping distance.
- the target feed finally positioned by the swiping in the feed streaming is determined based on the page swiping distance and the page display sizes of respective feeds.
- the feed N will be displayed at the bottom of the screen when the feed streaming stops moving, and then the feed N is considered as the target feed finally positioned.
- the content of the feed N is obtained and displayed.
- page position information of respective feeds is determined based on the display parameters of respective feeds.
- the actual distance is obtained based on the page position information of the operation feed of the swiping and the page position information of the target feed finally positioned.
- the offset between the page swiping distance and the actual distance is obtained.
- the feed streaming is moved based on the offset, in response to the offset being less than the threshold.
- FIGS. 1-8 are shown in sequence as indicated by the arrows, they are not necessarily executed in the order indicated by the arrows. Unless explicitly stated in the disclosure, there is no strict order in which these steps can be executed, and these steps can be executed in any other order. Moreover, at least part of the steps in FIGS. 1-8 may include more sub steps or multiple stages, these sub steps or stages are not necessarily executed at the same time, but may also be executed at different times. These sub steps or stages are not necessarily executed in sequence, instead, they may be executed alternately with at least part of the other steps or sub steps or stages in other steps.
- FIG. 9 is a block diagram of an apparatus 900 for obtaining content according to an embodiment. As illustrated in FIG. 9 , the apparatus includes a swiping speed obtaining module 901 , an object determining module 902 and a content displaying module 903 .
- the swiping speed obtaining module 901 is configured to obtain a swiping speed of a swiping on a page in response to detecting an end of the swiping.
- the object determining module 902 is configured to determine an object positioned by the swiping in the page based on the swiping speed and page parameters of the page.
- the content displaying module 903 is configured to obtain and display content of the object positioned by the swiping.
- the object determining module 902 includes a page swiping distance determining unit and an object determining unit.
- the page swiping distance determining unit is configured to determine a page swiping distance based on the swiping speed and the page parameters.
- the object determining unit is configured to determine the object positioned by the swiping in the page based on the page swiping distance and the page parameters.
- the page parameters include display parameters of a plurality of objects in the page
- the display parameters are obtained at first loading of the plurality of objects
- the object determining unit is configured to: determine page display sizes of respective objects based on the display parameters of respective objects; and determine the object positioned by the swiping in the page based on the page swiping distance and the page display sizes of respective objects.
- the page parameters include screen position information of an initial object of the swiping in a screen, the initial object is in the page, and the page swiping distance determining unit includes an initial page swiping distance determining unit, an inertial swiping distance determining unit and a page swiping distance obtaining unit.
- the initial page swiping distance determining unit is configured to determine an initial page swiping distance of the page during the swiping, based on the screen position information of the initial object.
- the inertial swiping distance determining unit is configured to determine an inertial swiping distance of the page after the end of the swiping, based on the swiping speed.
- the page swiping distance obtaining unit is configured to obtain a sum of the initial page swiping distance and the inertial swiping distance as the page swiping distance.
- the initial page swiping distance determining unit is configured to: obtain a first screen position of the initial object in response to a trigger of the swiping; obtain a second screen position of the initial object in response to the end of the swiping; and determine a swiping distance of the initial object in a page swiping direction based on the first screen position and the second screen position, as the initial page swiping distance.
- the inertial swiping distance determining unit is configured to determine the inertial swiping distance of the page after the end of the swiping, based on the swiping speed and a known inertia constant.
- the apparatus 900 further includes an inertial constant determining module.
- the inertial constant determining module is configured to obtain a plurality of historical inertia constants and obtain the inertia constant by processing the plurality of historical inertia constants.
- the swiping speed obtaining module is configured to obtain a distance travelled by the swiping per unit time as the swiping speed.
- the apparatus 900 further includes a page position determining module, a distance obtaining module, an offset obtaining module and a moving module.
- the page position determining module is configured to determine page positions of respective objects based on the display parameters of respective objects.
- the distance obtaining module is configured to obtain a distance based on the page position of the initial object of the swiping and the page position of the object positioned by the swiping.
- the offset obtaining module is configured to obtain an offset between the page swiping distance and the distance determined by the distance obtaining module.
- the moving module is configured to move the page based on the offset, in response to the offset being less than a threshold.
- FIG. 10 is a block diagram of an apparatus for obtaining content according to another embodiment. As illustrated in FIG. 10 , the apparatus 10000 includes one or more processors 110000 and one or more memory 12000 .
- the one or more memory 12000 is configured to store instructions executable by the one or more processors 11000 .
- the one or more processors 11000 are configured to executes the instructions, so as to: obtain a swiping speed of a swiping on a page in response to detecting an end of the swiping; determine an object positioned by the swiping in the page based on the swiping speed and page parameters of the page; and display content of the object.
- the one or more processors 11000 are configured to: determine a page swiping distance based on the swiping speed and the page parameters; and determine the object positioned by the swiping in the page based on the page swiping distance and the page parameters.
- the page parameters include display parameters of a plurality of objects in the page
- the display parameters are obtained at first loading of the plurality of objects
- the one or more processors 11000 are configured to: determine page display sizes of respective objects based on the display parameters of respective objects; and determine the object positioned by the swiping in the page based on the page swiping distance and the page display sizes of respective objects.
- the page parameters include screen position information of an initial object of the swiping in a screen, the initial object is in the page, and the one or more processors 11000 are configured to: determine an initial page swiping distance of the page during the swiping, based on the screen position information of the initial object; determine an inertial swiping distance of the page after the end of the swiping, based on the swiping speed; and determine the page swiping distance as a sum of the initial page swiping distance and the inertial swiping distance.
- the one or more processors 11000 are configured to: obtain a first screen position of the initial object in response to a trigger of the swiping; obtain a second screen position of the initial object in response to the end of the swiping; and determine a swiping distance of the initial object in a page swiping direction based on the first screen position and the second screen position, as the initial page swiping distance.
- the one or more processors 11000 are configured to: determine the inertial swiping distance of the page after the end of the swiping, based on the swiping speed and a known inertia constant.
- the one or more processors 11000 are further configured to: obtain a plurality of historical inertia constants and obtain the inertia constant by processing the plurality of historical inertia constants.
- the one or more processors 11000 are configured to determine the swiping speed based on a distance travelled by the swiping per unit time.
- the one or more processors 11000 are further configured to: determine page positions of respective objects based on the display parameters of respective objects; obtain a distance based on the page position of an initial object of the swiping and the page position of the object positioned by the swiping; obtain an offset between the page swiping distance and the distance; and move the page based on the offset, in response to the offset being less than a threshold.
- FIG. 11 is a block diagram of an electronic device 1000 according to an embodiment.
- the electronic device 1000 may be a mobile phone, a computer, a digital broadcasting terminal, a messaging device, a game console, a tablet device, a medical device, a fitness device, a personal digital assistant, etc.
- the electronic device 1000 may include one or more of the following components: a processing component 1002 , a memory 1004 , a power component 1006 , a multimedia component 1008 , an audio component 1010 , an input/output (I/O) interface 1012 , a sensor component 1014 and a communication component 1016 .
- the processing component 1002 generally controls the overall operations of the electronic device 1000 , such as the operations associated with display, telephone calls, data communications, camera operations, and recording operations.
- the processing component 1002 may include one or more processors 1020 to execute instructions to complete all or part of the steps of the aforementioned method.
- the processing component 1002 may include one or more modules to facilitate the interaction between the processing component 1002 and other components.
- the processing component 1002 may include a multimedia module to facilitate the interaction between the multimedia component 1008 and the processing component 1002 .
- the memory 1004 is configured to store various types of data to support the operation of the electronic device 1000 . Examples of these data include instructions for any application or method operating on the device 1000 , contact data, phone book data, messages, pictures, videos, etc.
- the memory 1004 may be implemented by any type of volatile or non-volatile storage devices or their combination, such as a static random access memory (SRAM), an erasable Programmable read-only memory (EPROM), a programmable read-only memory (PROM), a read-only memory (ROM), a magnetic memory, a flash memory, a magnetic disk or an optical disk.
- SRAM static random access memory
- EPROM erasable Programmable read-only memory
- PROM programmable read-only memory
- ROM read-only memory
- magnetic memory a magnetic memory
- flash memory a magnetic disk or an optical disk.
- the power component 1006 provides power to various components of the electronic device 1000 .
- the power component 1006 may include a power management system, one or more power supplies, and other components associated with generating, managing, and distributing power for the device 1000 .
- the multimedia component 1008 includes a screen that provides an output interface between the device 1000 and the user.
- the screen may include a liquid crystal display (LCD) and a touch panel (TP). If the screen includes a touch panel, the screen may be implemented as a touch screen to receive input signals from the user.
- the touch panel includes one or more touch sensors to sense touching, swiping, and gestures on the touch panel. The touch sensor may not only sense the boundary of a touch or slide action, but also detect the duration and pressure related to the touch or slide operation.
- the multimedia component 1008 includes a front camera and/or a rear camera. When the electronic device 1000 is in an operation mode, such as a shooting mode or a video mode, the front camera and/or the rear camera may receive external multimedia data. Each front camera and rear camera may be a fixed optical lens system or have focal length and optical zoom capabilities.
- the audio component 1010 is configured to output and/or input audio signals.
- the audio component 1010 includes a microphone (MIC).
- the microphone is configured to receive external audio signals.
- the received audio signal may be further stored in the memory 1004 or transmitted via the communication component 1016 .
- the audio component 1010 further includes a speaker for outputting audio signals.
- the I/O interface 1012 provides an interface between the processing component 1002 and a peripheral interface module.
- the peripheral interface module may be a keyboard, a click wheel, a button, and the like. These buttons may include but are not limited to: a home button, a volume button, a start button, and a lock button.
- the sensor component 1014 includes one or more sensors for providing the electronic device 1000 with various aspects of status assessment.
- the sensor component 1014 may detect the on/off status of the electronic device 1000 and the relative positioning of components.
- the component is a display and a keypad of the electronic device 1000 .
- the sensor component 1014 may also detect the position change of the electronic device 1000 or a component thereof. The presence or absence of a contact between the user and the electronic device 1000 , the orientation or acceleration/deceleration of the electronic device 1000 , and temperature change of the electronic device 1000 .
- the sensor component 1014 may include a proximity sensor configured to detect the presence of nearby objects when there is no physical contact.
- the sensor component 1014 may also include a light sensor, such as a CMOS or CCD image sensor, for use in imaging applications.
- the sensor component 1014 may also include an acceleration sensor, a gyroscope sensor, a magnetic sensor, a pressure sensor or a temperature sensor.
- the communication component 1016 is configured to facilitate wired or wireless communication between the electronic device 1000 and other devices.
- the electronic device 1000 may access a wireless network based on a communication standard, such as WiFi, 2G, or 3G, or a combination thereof.
- the communication component 1016 receives a broadcast signal or broadcast related information from an external broadcast management system via a broadcast channel.
- the communication component 1016 further includes a near field communication (NFC) module to facilitate short-range communication.
- the NFC module may be implemented based on radio frequency identification (RFID) technology, infrared data association (IrDA) technology, ultra-wideband (UWB) technology, Bluetooth (BT) technology and other technologies.
- RFID radio frequency identification
- IrDA infrared data association
- UWB ultra-wideband
- Bluetooth Bluetooth
- the electronic device 1000 may be implemented by one or more application specific integrated circuits (ASIC), digital signal processors (DSP), digital signal processing devices (DSPD), programmable logic devices (PLD), field programmable implemented by a gate array (FPGA), controllers, microcontrollers, microprocessors, or other electronic components, which are configured to implement the above voice control method.
- ASIC application specific integrated circuits
- DSP digital signal processors
- DSPD digital signal processing devices
- PLD programmable logic devices
- FPGA field programmable implemented by a gate array
- controllers microcontrollers, microprocessors, or other electronic components, which are configured to implement the above voice control method.
- non-transitory computer-readable storage medium including instructions, such as the memory 1004 including instructions, which may be executed by the processor 1020 of the electronic device 1000 to complete the aforementioned method.
- the non-transitory computer-readable storage medium may be a ROM, a random access memory (RAM), a CD-ROM, a magnetic tape, a floppy disk, an optical data storage device, etc.
Landscapes
- Engineering & Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
Description
where, s is the inertial swiping distance, v is the swiping speed, and A is the inertia constant.
offset=(initial page swiping distance+inertial swiping distance)−actual distance.
Claims (16)
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010254542.7A CN111427504A (en) | 2020-04-02 | 2020-04-02 | Content acquisition method and device, electronic equipment and storage medium |
CN202010254542.7 | 2020-04-02 |
Publications (2)
Publication Number | Publication Date |
---|---|
US20210311622A1 US20210311622A1 (en) | 2021-10-07 |
US11474689B2 true US11474689B2 (en) | 2022-10-18 |
Family
ID=71550588
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US17/136,667 Active 2041-01-07 US11474689B2 (en) | 2020-04-02 | 2020-12-29 | Method and apparatus for obtaining content |
Country Status (2)
Country | Link |
---|---|
US (1) | US11474689B2 (en) |
CN (1) | CN111427504A (en) |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113391743B (en) * | 2021-04-29 | 2022-05-20 | 荣耀终端有限公司 | A display method and electronic device |
CN113238709A (en) * | 2021-06-02 | 2021-08-10 | 北京字节跳动网络技术有限公司 | Control method, device and equipment based on touch pad |
CN113867591A (en) * | 2021-10-09 | 2021-12-31 | 拉扎斯网络科技(上海)有限公司 | Control method and device for displaying second floor page, storage medium and electronic equipment |
CN113986067A (en) * | 2021-10-29 | 2022-01-28 | 维沃移动通信有限公司 | Object control method, device, equipment and storage medium |
CN115237325A (en) * | 2022-07-12 | 2022-10-25 | 抖音视界(北京)有限公司 | Page content updating method and device, computer equipment and storage medium |
Citations (35)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100235794A1 (en) * | 2009-03-16 | 2010-09-16 | Bas Ording | Accelerated Scrolling for a Multifunction Device |
US20100295781A1 (en) * | 2009-05-22 | 2010-11-25 | Rachid Alameh | Electronic Device with Sensing Assembly and Method for Interpreting Consecutive Gestures |
US20110074699A1 (en) * | 2009-09-25 | 2011-03-31 | Jason Robert Marr | Device, Method, and Graphical User Interface for Scrolling a Multi-Section Document |
US20120174005A1 (en) * | 2010-12-31 | 2012-07-05 | Microsoft Corporation | Content-based snap point |
US20120218310A1 (en) * | 2011-02-24 | 2012-08-30 | Kyocera Corporation | Electronic device, display control method, and storage medium storing display control program |
US20120218203A1 (en) * | 2011-02-10 | 2012-08-30 | Kanki Noriyoshi | Touch drawing display apparatus and operation method thereof, image display apparatus allowing touch-input, and controller for the display apparatus |
US20130111396A1 (en) * | 2011-10-31 | 2013-05-02 | Microsoft Corporation | Exposing inertial snap points |
US20130179830A1 (en) * | 2012-01-09 | 2013-07-11 | Samsung Electronics Co., Ltd. | Graphical user interface, display apparatus and control method thereof |
US20130207909A1 (en) * | 2012-02-09 | 2013-08-15 | Kabushiki Kaisha Square Enix (Also Trading As Square Enix Co., Ltd.) | Scrolling screen apparatus, method for scrolling screen, and game apparatus |
CN103729124A (en) | 2012-10-12 | 2014-04-16 | 腾讯科技(深圳)有限公司 | Control method and system for slide list |
US20140181730A1 (en) * | 2012-12-21 | 2014-06-26 | Orange | Fragmented scrolling of a page |
US20140176455A1 (en) * | 2012-12-26 | 2014-06-26 | Panasonic Corporation | Touch panel device and method of controlling a touch panel device |
US20150058761A1 (en) * | 2013-08-26 | 2015-02-26 | Lenovo (Beijing) Co., Ltd. | Information processing method and electronic device |
US8971572B1 (en) * | 2011-08-12 | 2015-03-03 | The Research Foundation For The State University Of New York | Hand pointing estimation for human computer interaction |
US20160070429A1 (en) * | 2014-09-04 | 2016-03-10 | Home Box Office, Inc. | Snap points including ranges |
US20160124597A1 (en) * | 2013-06-26 | 2016-05-05 | Kyocera Corporation | Electronic device and display control method |
US20160179322A1 (en) * | 2013-08-28 | 2016-06-23 | Kyocera Corporation | Electronic device and method for controlling electronic device |
US20160357382A1 (en) * | 2015-06-07 | 2016-12-08 | Apple Inc. | Intelligent Scrolling of Electronic Document |
US20170099431A1 (en) * | 2015-10-06 | 2017-04-06 | Canon Kabushiki Kaisha | Image processing apparatus, electronic apparatus, display processing apparatus, and method for controlling the same |
CN106933481A (en) | 2015-12-29 | 2017-07-07 | 苏宁云商集团股份有限公司 | A kind of screen scroll method and device |
CN107168628A (en) | 2017-04-25 | 2017-09-15 | 福建星网视易信息系统有限公司 | A kind of screen sliding display methods and device |
US20170285861A1 (en) * | 2016-03-31 | 2017-10-05 | Rovi Guides, Inc. | Systems and methods for reducing jitter using a touch screen |
US9851896B2 (en) * | 2013-12-17 | 2017-12-26 | Google Inc. | Edge swiping gesture for home navigation |
CN109086374A (en) | 2018-07-23 | 2018-12-25 | 北京达佳互联信息技术有限公司 | Webpage loading display method, device, terminal device and readable storage medium storing program for executing |
CN109388317A (en) | 2018-09-30 | 2019-02-26 | 武汉斗鱼网络科技有限公司 | A kind of picture loading method, terminal device and storage medium |
CN109669740A (en) | 2018-09-25 | 2019-04-23 | 平安科技(深圳)有限公司 | List loading method, device, equipment and computer readable storage medium |
US20190324618A1 (en) * | 2013-06-11 | 2019-10-24 | Sony Corporation | Apparatus, method, computer-readable storage medium, and smartphone for causing scrolling of content in response to touch operations |
CN110611829A (en) | 2019-09-11 | 2019-12-24 | 北京达佳互联信息技术有限公司 | Video preloading method and device, electronic equipment and storage medium |
US20200058270A1 (en) * | 2017-04-28 | 2020-02-20 | Huawei Technologies Co., Ltd. | Bullet screen display method and electronic device |
US10585577B2 (en) * | 2016-06-20 | 2020-03-10 | Kyocera Document Solutions, Inc. | Display device and computer-readable non-transitory recording medium recording display control program |
US10606461B1 (en) * | 2012-07-26 | 2020-03-31 | Google Llc | Snapping a pointing-indicator to a scene boundary of a video |
US10642365B2 (en) * | 2014-09-09 | 2020-05-05 | Microsoft Technology Licensing, Llc | Parametric inertia and APIs |
US20200142554A1 (en) * | 2017-06-27 | 2020-05-07 | Microsoft Technology Licensing, Llc | Page navigation in desktop launcher |
US10891044B1 (en) * | 2016-10-25 | 2021-01-12 | Twitter, Inc. | Automatic positioning of content items in a scrolling display for optimal viewing of the items |
US20210149534A1 (en) * | 2018-05-21 | 2021-05-20 | Huawei Technologies Co., Ltd. | Display Control Method and Terminal |
-
2020
- 2020-04-02 CN CN202010254542.7A patent/CN111427504A/en active Pending
- 2020-12-29 US US17/136,667 patent/US11474689B2/en active Active
Patent Citations (37)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100235794A1 (en) * | 2009-03-16 | 2010-09-16 | Bas Ording | Accelerated Scrolling for a Multifunction Device |
US20100295781A1 (en) * | 2009-05-22 | 2010-11-25 | Rachid Alameh | Electronic Device with Sensing Assembly and Method for Interpreting Consecutive Gestures |
US20110074699A1 (en) * | 2009-09-25 | 2011-03-31 | Jason Robert Marr | Device, Method, and Graphical User Interface for Scrolling a Multi-Section Document |
US20120174005A1 (en) * | 2010-12-31 | 2012-07-05 | Microsoft Corporation | Content-based snap point |
US20120218203A1 (en) * | 2011-02-10 | 2012-08-30 | Kanki Noriyoshi | Touch drawing display apparatus and operation method thereof, image display apparatus allowing touch-input, and controller for the display apparatus |
US20120218310A1 (en) * | 2011-02-24 | 2012-08-30 | Kyocera Corporation | Electronic device, display control method, and storage medium storing display control program |
US8971572B1 (en) * | 2011-08-12 | 2015-03-03 | The Research Foundation For The State University Of New York | Hand pointing estimation for human computer interaction |
US20130111396A1 (en) * | 2011-10-31 | 2013-05-02 | Microsoft Corporation | Exposing inertial snap points |
US20130179830A1 (en) * | 2012-01-09 | 2013-07-11 | Samsung Electronics Co., Ltd. | Graphical user interface, display apparatus and control method thereof |
US20130207909A1 (en) * | 2012-02-09 | 2013-08-15 | Kabushiki Kaisha Square Enix (Also Trading As Square Enix Co., Ltd.) | Scrolling screen apparatus, method for scrolling screen, and game apparatus |
US10606461B1 (en) * | 2012-07-26 | 2020-03-31 | Google Llc | Snapping a pointing-indicator to a scene boundary of a video |
CN103729124A (en) | 2012-10-12 | 2014-04-16 | 腾讯科技(深圳)有限公司 | Control method and system for slide list |
US20140181730A1 (en) * | 2012-12-21 | 2014-06-26 | Orange | Fragmented scrolling of a page |
US9880726B2 (en) * | 2012-12-21 | 2018-01-30 | Orange | Fragmented scrolling of a page |
US20140176455A1 (en) * | 2012-12-26 | 2014-06-26 | Panasonic Corporation | Touch panel device and method of controlling a touch panel device |
US20190324618A1 (en) * | 2013-06-11 | 2019-10-24 | Sony Corporation | Apparatus, method, computer-readable storage medium, and smartphone for causing scrolling of content in response to touch operations |
US20160124597A1 (en) * | 2013-06-26 | 2016-05-05 | Kyocera Corporation | Electronic device and display control method |
US20150058761A1 (en) * | 2013-08-26 | 2015-02-26 | Lenovo (Beijing) Co., Ltd. | Information processing method and electronic device |
US20160179322A1 (en) * | 2013-08-28 | 2016-06-23 | Kyocera Corporation | Electronic device and method for controlling electronic device |
US9851896B2 (en) * | 2013-12-17 | 2017-12-26 | Google Inc. | Edge swiping gesture for home navigation |
US10019146B2 (en) * | 2014-09-04 | 2018-07-10 | Home Box Office, Inc. | Snap points including ranges |
US20160070429A1 (en) * | 2014-09-04 | 2016-03-10 | Home Box Office, Inc. | Snap points including ranges |
US10642365B2 (en) * | 2014-09-09 | 2020-05-05 | Microsoft Technology Licensing, Llc | Parametric inertia and APIs |
US20160357382A1 (en) * | 2015-06-07 | 2016-12-08 | Apple Inc. | Intelligent Scrolling of Electronic Document |
US20170099431A1 (en) * | 2015-10-06 | 2017-04-06 | Canon Kabushiki Kaisha | Image processing apparatus, electronic apparatus, display processing apparatus, and method for controlling the same |
CN106933481A (en) | 2015-12-29 | 2017-07-07 | 苏宁云商集团股份有限公司 | A kind of screen scroll method and device |
US20170285861A1 (en) * | 2016-03-31 | 2017-10-05 | Rovi Guides, Inc. | Systems and methods for reducing jitter using a touch screen |
US10585577B2 (en) * | 2016-06-20 | 2020-03-10 | Kyocera Document Solutions, Inc. | Display device and computer-readable non-transitory recording medium recording display control program |
US10891044B1 (en) * | 2016-10-25 | 2021-01-12 | Twitter, Inc. | Automatic positioning of content items in a scrolling display for optimal viewing of the items |
CN107168628A (en) | 2017-04-25 | 2017-09-15 | 福建星网视易信息系统有限公司 | A kind of screen sliding display methods and device |
US20200058270A1 (en) * | 2017-04-28 | 2020-02-20 | Huawei Technologies Co., Ltd. | Bullet screen display method and electronic device |
US20200142554A1 (en) * | 2017-06-27 | 2020-05-07 | Microsoft Technology Licensing, Llc | Page navigation in desktop launcher |
US20210149534A1 (en) * | 2018-05-21 | 2021-05-20 | Huawei Technologies Co., Ltd. | Display Control Method and Terminal |
CN109086374A (en) | 2018-07-23 | 2018-12-25 | 北京达佳互联信息技术有限公司 | Webpage loading display method, device, terminal device and readable storage medium storing program for executing |
CN109669740A (en) | 2018-09-25 | 2019-04-23 | 平安科技(深圳)有限公司 | List loading method, device, equipment and computer readable storage medium |
CN109388317A (en) | 2018-09-30 | 2019-02-26 | 武汉斗鱼网络科技有限公司 | A kind of picture loading method, terminal device and storage medium |
CN110611829A (en) | 2019-09-11 | 2019-12-24 | 北京达佳互联信息技术有限公司 | Video preloading method and device, electronic equipment and storage medium |
Non-Patent Citations (2)
Title |
---|
Chinese Office Action dated Jan. 25, 2021 for Application No. 202010254542.7, 17 pages. |
Indian Office Action dated Jan. 3, 2022 for Application No. 202014057199, 6 pages. |
Also Published As
Publication number | Publication date |
---|---|
US20210311622A1 (en) | 2021-10-07 |
CN111427504A (en) | 2020-07-17 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11474689B2 (en) | Method and apparatus for obtaining content | |
EP3460647B1 (en) | Method for controlling a screen, device and storage medium | |
CN106294609B (en) | Page loading method and device | |
CN106537319A (en) | Screen-splitting display method and device | |
US10331231B2 (en) | Mobile terminal and method for determining scrolling speed | |
EP3885885A1 (en) | Method, apparatus and storage medium for displaying application interface | |
EP3454198A1 (en) | Method and apparatus for controlling application | |
US20180039403A1 (en) | Terminal control method, terminal, and storage medium | |
US20200183549A1 (en) | Application icon moving method and apparatus, terminal and storage medium | |
CN103927101B (en) | The method and apparatus of operational controls | |
US11644942B2 (en) | Method and device for displaying application, and storage medium | |
EP3890337A1 (en) | Method and device for loading electronic program guide (epg) | |
CN107694087B (en) | Information processing method and terminal equipment | |
US11157085B2 (en) | Method and apparatus for switching display mode, mobile terminal and storage medium | |
US11513679B2 (en) | Method and apparatus for processing touch signal, and medium | |
US20220137763A1 (en) | Small-screen window display method, device and storage medium | |
EP3816778B1 (en) | Punch-hole screen display method and apparatus | |
EP3889913B1 (en) | Animation generation | |
CN107870932B (en) | User word stock optimization method and device and electronic equipment | |
CN107870931B (en) | Method and device for optimizing user lexicon and electronic equipment | |
US20210333988A1 (en) | Method and device for switching interface, touch terminal, and storage medium | |
CN108469926B (en) | Equipment control method, device and equipment | |
CN105894558B (en) | GIF image rendering method and device | |
CN119149148A (en) | Processing method, device and medium for form data | |
CN105373333A (en) | Touch response method and apparatus |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
AS | Assignment |
Owner name: BEIJING DAJIA INTERNET INFORMATION TECHNOLOGY CO., LTD., CHINA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:QU, MEIKAI;REEL/FRAME:054830/0421 Effective date: 20201105 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |