[go: up one dir, main page]

 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (35)

Search Parameters:
Keywords = fuzzy reward

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
17 pages, 1577 KiB  
Article
Intelligent Traffic Control Decision-Making Based on Type-2 Fuzzy and Reinforcement Learning
by Yunrui Bi, Qinglin Ding, Yijun Du, Di Liu and Shuaihang Ren
Electronics 2024, 13(19), 3894; https://doi.org/10.3390/electronics13193894 - 1 Oct 2024
Viewed by 580
Abstract
Intelligent traffic control decision-making has long been a crucial issue for improving the efficiency and safety of the intelligent transportation system. The deficiencies of the Type-1 fuzzy traffic control system in dealing with uncertainty have led to a reduced ability to address traffic [...] Read more.
Intelligent traffic control decision-making has long been a crucial issue for improving the efficiency and safety of the intelligent transportation system. The deficiencies of the Type-1 fuzzy traffic control system in dealing with uncertainty have led to a reduced ability to address traffic congestion. Therefore, this paper proposes a Type-2 fuzzy controller for a single intersection. Based on real-time traffic flow information, the green timing of each phase is dynamically determined to achieve the minimum average vehicle delay. Additionally, in traffic light control, various factors (such as vehicle delay and queue length) need to be balanced to define the appropriate reward. Improper reward design may fail to guide the Deep Q-Network algorithm to learn the optimal strategy. To address these issues, this paper proposes a deep reinforcement learning traffic control strategy combined with Type-2 fuzzy control. The output action of the Type-2 fuzzy control system replaces the action of selecting the maximum output Q-value of the target network in the DQN algorithm, reducing the error caused by the use of the max operation of the target network. This approach improves the online learning rate of the agent and increases the reward value of the signal control action. The simulation results using the Simulation of Urban MObility platform show that the traffic signal optimization control proposed in this paper has achieved significant improvement in traffic flow optimization and congestion alleviation, which can effectively improve the traffic efficiency in front of the signal light and improve the overall operation level of traffic flow. Full article
(This article belongs to the Special Issue Smart Vehicles and Smart Transportation Research Trends)
Show Figures

Figure 1

Figure 1
<p>Single intersection signal light control model.</p>
Full article ">Figure 2
<p>There are three multiple panels in the process of converting the traffic state into the input matrix.</p>
Full article ">Figure 3
<p>Four-phase signal diagram.</p>
Full article ">Figure 4
<p>Traffic decision principle diagram based on Type2-FDQN algorithm.</p>
Full article ">Figure 5
<p>The workflow of the traffic decision-making process based on the Type-2-FDQN algorithm.</p>
Full article ">Figure 6
<p>SUMO simulation single-intersection simulation environment.</p>
Full article ">Figure 7
<p>Software simulation process.</p>
Full article ">Figure 8
<p>Trend chart of average cumulative reward value.</p>
Full article ">Figure 9
<p>Average queue length of the vehicle.</p>
Full article ">Figure 10
<p>Average speed of vehicle.</p>
Full article ">Figure 11
<p>Average waiting time of the vehicle.</p>
Full article ">Figure 12
<p>Total waiting time of vehicle.</p>
Full article ">Figure 13
<p>Average vehicle speed under different traffic volumes.</p>
Full article ">Figure 14
<p>Average vehicle queue length under different traffic volumes.</p>
Full article ">
28 pages, 6881 KiB  
Article
Engagement Analysis Using Electroencephalography Signals in Games for Hand Rehabilitation with Dynamic and Random Difficulty Adjustments
by Raúl Daniel García-Ramón, Ericka Janet Rechy-Ramirez, Luz María Alonso-Valerdi and Antonio Marin-Hernandez
Appl. Sci. 2024, 14(18), 8464; https://doi.org/10.3390/app14188464 - 20 Sep 2024
Viewed by 706
Abstract
Background: Traditional physical rehabilitation involves participants performing repetitive body movements with the assistance of physiotherapists. Owing to the exercises’ monotonous nature and lack of reward, participants may become disinterested and cease their recovery. Games could be used as tools to engage participants in [...] Read more.
Background: Traditional physical rehabilitation involves participants performing repetitive body movements with the assistance of physiotherapists. Owing to the exercises’ monotonous nature and lack of reward, participants may become disinterested and cease their recovery. Games could be used as tools to engage participants in the rehabilitation process. Consequently, participants could perform rehabilitation exercises while playing the game, receiving rewards from the experience. Maintaining the players’ engagement requires regularly adjusting the game difficulty. The players’ engagement can be measured using questionnaires and biosignals (e.g., electroencephalography signals—EEG). This study aims to determine whether there is a significant difference in players’ engagement between two game modes with different game difficulty adjustments: non-tailored and tailored modes. Methods: We implemented two game modes which were controlled using hand movements. The features of the game rewards (position and size) were changed in the game scene; hence, the game difficulty could be modified. The non-tailored mode set the features of rewards in the game scene randomly. Conversely, the tailored mode set the features of rewards in the game scene based on the participants’ range of motion using fuzzy logic. Consequently, the game difficulty was adjusted dynamically. Additionally, engagement was computed from 53 healthy participants in both game modes using two EEG sensors: Bitalino Revolution and Unicorn. Specifically, the theta (θ) and alpha (α) bands from the frontal and parietal lobes were computed from the EEG data. A questionnaire was applied to participants after finishing playing both game modes to collect their impressions on the following: their favorite game mode, the game mode that was the easiest to play, the game mode that was the least frustrating to play, the game mode that was the least boring to play, the game mode that was the most entertaining to play, and the game mode that had the fastest game response time. Results: The non-tailored game mode reported the following means of engagement: 6.297 ± 11.274 using the Unicorn sensor, and 3.616 ± 0.771 using the Bitalino sensor. The tailored game mode reported the following means of engagement: 4.408 ± 6.243 using the Unicorn sensor, and 3.619 ± 0.551 using Bitalino. The non-tailored mode reported the highest mean engagement (6.297) when the Unicorn sensor was used to collect EEG signals. Most participants selected the non-tailored game mode as their favorite, and the most entertaining mode, irrespective of the EEG sensor. Conversely, most participants chose the tailored game mode as the easiest, and the least frustrating mode to play, irrespective of the EEG sensor. Conclusions: A Wilcoxon-Signed-Rank test revealed that there was only a significant difference in engagement between game modes when the EEG signal was collected via the Unicorn sensor (p value = 0.04054). Fisher’s exact tests showed significant associations between the game modes (non-tailored, tailored) and the following players’ variables: ease of play using the Unicorn sensor (p value = 0.009341), and frustration using Unicorn sensor (p value = 0.0466). Full article
(This article belongs to the Special Issue Serious Games and Extended Reality in Healthcare)
Show Figures

Figure 1

Figure 1
<p>Game stages: (<b>a</b>) Game stage 1. (<b>b</b>) Game stage 2.</p>
Full article ">Figure 2
<p>Hand movements used to play the game stages.</p>
Full article ">Figure 3
<p>Fuzzy input sets based on the ROMs that were used to compute the box position: (<b>a</b>) Fuzzy input sets for ulnar movement. (<b>b</b>) Fuzzy input sets for the extension movement.</p>
Full article ">Figure 4
<p>Fuzzy input sets based on the ROMs that were used to compute the box size: (<b>a</b>) Fuzzy input sets for radial movement. (<b>b</b>) Fuzzy input sets for the flexion movement.</p>
Full article ">Figure 5
<p>Fuzzy output sets for the box position: (<b>a</b>) Fuzzy output sets for the X axis. (<b>b</b>) Fuzzy output sets for the Y axis.</p>
Full article ">Figure 6
<p>Fuzzy output sets for the size position: (<b>a</b>) Fuzzy output sets for the X axis. (<b>b</b>) Fuzzy output sets for the Y axis.</p>
Full article ">Figure 7
<p>Experimental setting using the Bitalino sensor to collect the EEG signal.</p>
Full article ">Figure 8
<p>Experimental setting using the Unicorn sensor to collect the EEG signal.</p>
Full article ">Figure 9
<p>Participants’ favorite game mode in experimental group 1—Unicorn sensor.</p>
Full article ">Figure 10
<p>Participants’ opinions on the ease of playing the game modes using the Unicorn sensor.</p>
Full article ">Figure 11
<p>Participants’ opinions on frustration in game modes using the Unicorn sensor.</p>
Full article ">Figure 12
<p>Participants’ opinions on boredom in game modes using the Unicorn sensor.</p>
Full article ">Figure 13
<p>Participants’ opinions on game response time to commands in game modes using the Unicorn sensor.</p>
Full article ">Figure 14
<p>Participants’ opinions on entertainment in game modes using the Unicorn sensor.</p>
Full article ">Figure 15
<p>Participants’ favorite game mode in experimental group 2—Bitalino sensor.</p>
Full article ">Figure 16
<p>Participants’ opinions on frustration in game modes using the Bitalino sensor.</p>
Full article ">Figure 17
<p>Participants’ opinions on the ease of playing in game modes using the Bitalino sensor.</p>
Full article ">Figure 18
<p>Participants’ opinions on boredom in game modes using the Bitalino sensor.</p>
Full article ">Figure 19
<p>Participants’ opinions on game response time to the commands in game modes using the Bitalino sensor.</p>
Full article ">Figure 20
<p>Participants’ opinions on entertainment in game modes using the Bitalino sensor.</p>
Full article ">
26 pages, 3235 KiB  
Article
Traffic Signal Control with State-Optimizing Deep Reinforcement Learning and Fuzzy Logic
by Teerapun Meepokgit and Sumek Wisayataksin
Appl. Sci. 2024, 14(17), 7908; https://doi.org/10.3390/app14177908 - 5 Sep 2024
Viewed by 900
Abstract
Traffic lights are the most commonly used tool to manage urban traffic to reduce congestion and accidents. However, the poor management of traffic lights can result in further problems. Consequently, many studies on traffic light control have been conducted using deep reinforcement learning [...] Read more.
Traffic lights are the most commonly used tool to manage urban traffic to reduce congestion and accidents. However, the poor management of traffic lights can result in further problems. Consequently, many studies on traffic light control have been conducted using deep reinforcement learning in the past few years. In this study, we propose a traffic light control method in which a Deep Q-network with fuzzy logic is used to reduce waiting time while enhancing the efficiency of the method. Nevertheless, existing studies using the Deep Q-network may yield suboptimal results because of the reward function, leading to the system favoring straight vehicles, which results in left-turning vehicles waiting too long. Therefore, we modified the reward function to consider the waiting time in each lane. For the experiment, Simulation of Urban Mobility (SUMO) software version 1.18.0 was used for various environments and vehicle types. The results show that, when using the proposed method in a prototype environment, the average total waiting time could be reduced by 18.46% compared with the traffic light control method using a conventional Deep Q-network with fuzzy logic. Additionally, an ambulance prioritization system was implemented that significantly reduced the ambulance waiting time. In summary, the proposed method yielded better results in all environments. Full article
(This article belongs to the Special Issue Intelligent Transportation System Technologies and Applications)
Show Figures

Figure 1

Figure 1
<p>Reinforcement learning cycle [<a href="#B30-applsci-14-07908" class="html-bibr">30</a>].</p>
Full article ">Figure 2
<p>(<b>a</b>) The prototype intersection. (<b>b</b>) The three-lane environment on a single arm is called N-3. (<b>c</b>) The three-lane environment on two arms is called E-3 S-3. (<b>d</b>) The three-lane environment on three arms is called S-3 N-3 W-3. (<b>e</b>) Left-hand traffic environment.</p>
Full article ">Figure 3
<p>Ambulance detection distance in the ambulance prioritization system.</p>
Full article ">Figure 4
<p>(<b>a</b>) Overview of traffic at intersections. (<b>b</b>) Overview of the traffic situation on the southern side of the intersection, divided into cells. (<b>c</b>) Each car’s waiting time on the southern side of the intersection is divided into cells. (<b>d</b>) The stored state values located on the southern side of the intersection.</p>
Full article ">Figure 5
<p>The directions of traffic signals phase.</p>
Full article ">Figure 6
<p>Group of lanes.</p>
Full article ">Figure 7
<p>(<b>a</b>) Traffic generation over a single episode based on vehicle categories. (<b>b</b>) Traffic generation over a single episode based on vehicle direction. (<b>c</b>) Traffic generation over a single episode based on intersection direction.</p>
Full article ">Figure 8
<p>(<b>a</b>) The input fuzzy membership of the GP. (<b>b</b>) The input fuzzy membership of the RP.</p>
Full article ">Figure 9
<p>(<b>a</b>) The output fuzzy membership of the green duration. (<b>b</b>) Examples of results obtained from the defuzzification process.</p>
Full article ">Figure 10
<p>(<b>a</b>) Training result graph based on negative reward. (<b>b</b>) Training result graph based on average total waiting time. (<b>c</b>) Training result graph based on average total waiting time for vehicles to make a left turn. (<b>d</b>) Training result graph based on average total waiting time per car.</p>
Full article ">Figure 10 Cont.
<p>(<b>a</b>) Training result graph based on negative reward. (<b>b</b>) Training result graph based on average total waiting time. (<b>c</b>) Training result graph based on average total waiting time for vehicles to make a left turn. (<b>d</b>) Training result graph based on average total waiting time per car.</p>
Full article ">
19 pages, 3537 KiB  
Article
Integral-Valued Pythagorean Fuzzy-Set-Based Dyna Q+ Framework for Task Scheduling in Cloud Computing
by Bhargavi Krishnamurthy and Sajjan G. Shiva
Sensors 2024, 24(16), 5272; https://doi.org/10.3390/s24165272 - 14 Aug 2024
Viewed by 397
Abstract
Task scheduling is a critical challenge in cloud computing systems, greatly impacting their performance. Task scheduling is a nondeterministic polynomial time hard (NP-Hard) problem that complicates the search for nearly optimal solutions. Five major uncertainty parameters, i.e., security, traffic, workload, availability, and price, [...] Read more.
Task scheduling is a critical challenge in cloud computing systems, greatly impacting their performance. Task scheduling is a nondeterministic polynomial time hard (NP-Hard) problem that complicates the search for nearly optimal solutions. Five major uncertainty parameters, i.e., security, traffic, workload, availability, and price, influence task scheduling decisions. The primary rationale for selecting these uncertainty parameters lies in the challenge of accurately measuring their values, as empirical estimations often diverge from the actual values. The integral-valued Pythagorean fuzzy set (IVPFS) is a promising mathematical framework to deal with parametric uncertainties. The Dyna Q+ algorithm is the updated form of the Dyna Q agent designed specifically for dynamic computing environments by providing bonus rewards to non-exploited states. In this paper, the Dyna Q+ agent is enriched with the IVPFS mathematical framework to make intelligent task scheduling decisions. The performance of the proposed IVPFS Dyna Q+ task scheduler is tested using the CloudSim 3.3 simulator. The execution time is reduced by 90%, the makespan time is also reduced by 90%, the operation cost is below 50%, and the resource utilization rate is improved by 95%, all of these parameters meeting the desired standards or expectations. The results are also further validated using an expected value analysis methodology that confirms the good performance of the task scheduler. A better balance between exploration and exploitation through rigorous action-based learning is achieved by the Dyna Q+ agent. Full article
(This article belongs to the Special Issue AI Technology for Cybersecurity and IoT Applications)
Show Figures

Figure 1

Figure 1
<p>Proposed IVPFS-Dyna Q+ task scheduler.</p>
Full article ">Figure 2
<p>Different types of client workflows versus workflow execution time (ms).</p>
Full article ">Figure 3
<p>Different types of client workflows versus makespan time (ms).</p>
Full article ">Figure 4
<p>Different types of client workflows versus operation cost (ms).</p>
Full article ">Figure 5
<p>Different types of client workflows versus resource utilization rate.</p>
Full article ">Figure 6
<p>Different types of client workflows versus workflow execution time.</p>
Full article ">Figure 7
<p>Different types of client workflows versus makespan time.</p>
Full article ">Figure 8
<p>Different types of client workflows versus operation cost.</p>
Full article ">Figure 9
<p>Different types of client workflows versus resource utilization rate.</p>
Full article ">Figure 10
<p>Different types of client workflows versus task execution time (ms).</p>
Full article ">Figure 11
<p>Different types of client workflows versus makespan time (ms).</p>
Full article ">Figure 12
<p>Different types of client workflows versus operation cost (USD).</p>
Full article ">Figure 13
<p>Different types of client workflows versus resource utilization rate.</p>
Full article ">
17 pages, 1312 KiB  
Article
Optimization of Predefined-Time Agent-Scheduling Strategy Based on PPO
by Dingding Qi, Yingjun Zhao, Longyue Li and Zhanxiao Jia
Mathematics 2024, 12(15), 2387; https://doi.org/10.3390/math12152387 - 31 Jul 2024
Viewed by 604
Abstract
In this paper, we introduce an agent rescue scheduling approach grounded in proximal policy optimization, coupled with a singularity-free predefined-time control strategy. The primary objective of this methodology is to bolster the efficiency and precision of rescue missions. Firstly, we have designed an [...] Read more.
In this paper, we introduce an agent rescue scheduling approach grounded in proximal policy optimization, coupled with a singularity-free predefined-time control strategy. The primary objective of this methodology is to bolster the efficiency and precision of rescue missions. Firstly, we have designed an evaluation function closely related to the average flying distance of agents, which provides a quantitative benchmark for assessing different scheduling schemes and assists in optimizing the allocation of rescue resources. Secondly, we have developed a scheduling strategy optimization method using the Proximal Policy Optimization (PPO) algorithm. This method can automatically learn and adjust scheduling strategies to adapt to complex rescue environments and varying task demands. The evaluation function provides crucial feedback signals for the PPO algorithm, ensuring that the algorithm can precisely adjust the scheduling strategies to achieve optimal results. Thirdly, aiming to attain stability and precision in agent navigation to designated positions, we formulate a singularity-free predefined-time fuzzy adaptive tracking control strategy. This approach dynamically modulates control parameters in reaction to external disturbances and uncertainties, thus ensuring the precise arrival of agents at their destinations within the predefined time. Finally, to substantiate the validity of our proposed approach, we crafted a simulation environment in Python 3.7, engaging in a comparative analysis between the PPO and the other optimization method, Deep Q-network (DQN), utilizing the variation in reward values as the benchmark for evaluation. Full article
(This article belongs to the Section Engineering Mathematics)
Show Figures

Figure 1

Figure 1
<p>Agent scheduling results for Scenario 1. × represents the rescue area, • represents the agent. (<b>a</b>) The results of the random scheduling of agents. The average distance is 6.61. (<b>b</b>) The scheduling results after optimization using the PPO algorithm. The average distance is 5.40.</p>
Full article ">Figure 2
<p>Agent scheduling results for Scenario 2. × represents the rescue area, • represents the agent. (<b>a</b>) The results of the random scheduling of agents. The average distance is 4.88. (<b>b</b>) The scheduling results after optimization using the PPO algorithm. The average distance is 3.72.</p>
Full article ">Figure 3
<p>Comparison of the average return values between the PPO algorithm (in this paper) and the DQN algorithm (traditional).</p>
Full article ">Figure 4
<p>The trajectories of <span class="html-italic">y</span> and <math display="inline"><semantics> <msub> <mi>y</mi> <mi>d</mi> </msub> </semantics></math>.</p>
Full article ">Figure 5
<p>The trajectory of tracking error <math display="inline"><semantics> <msub> <mi>μ</mi> <mn>1</mn> </msub> </semantics></math>.</p>
Full article ">Figure 6
<p>The trajectory of control input <span class="html-italic">u</span>.</p>
Full article ">Figure 7
<p>The trajectories of adaptive parameters <math display="inline"><semantics> <mrow> <msub> <mover accent="true"> <mi mathvariant="sans-serif">Φ</mi> <mo stretchy="false">^</mo> </mover> <mi>i</mi> </msub> <mrow> <mo>(</mo> <mi>i</mi> <mo>=</mo> <mn>1</mn> <mo>,</mo> <mn>2</mn> <mo>)</mo> </mrow> </mrow> </semantics></math>.</p>
Full article ">
26 pages, 4316 KiB  
Article
Definition of Regulatory Targets for Electricity Default Rate in Brazil: Proposition of a Fuzzy Inference-Based Model
by Nivia Maria Celestino, Rodrigo Calili, Daniel Louzada and Maria Fatima Almeida
Energies 2024, 17(9), 2147; https://doi.org/10.3390/en17092147 - 30 Apr 2024
Viewed by 585
Abstract
The current electricity default rates in continental countries, such as Brazil, pose risks to the economic stability and investment capabilities of distribution utilities. This situation results in higher electricity tariffs for regular customers. From a regulatory perspective, the key issue regarding this challenge [...] Read more.
The current electricity default rates in continental countries, such as Brazil, pose risks to the economic stability and investment capabilities of distribution utilities. This situation results in higher electricity tariffs for regular customers. From a regulatory perspective, the key issue regarding this challenge is devising incentive mechanisms that reward distribution utilities for their operational and investment choices, aiming to mitigate or decrease electricity non-payment rates and avoid tariff increases for regular customers. Despite adhering to the principles of incentive regulation, the Brazilian Electricity Regulatory Agency (ANEEL) uses a methodological approach to define regulatory targets for electricity defaults tied to econometric models developed to determine targets to combat electricity non-technical losses (NTLs). This methodology has been widely criticized by electricity distribution utilities and academics because it includes many ad hoc steps and fails to consider the components that capture the specificities and heterogeneity of distribution utilities. This study proposes a fuzzy inference-based model for defining regulatory default targets built independently of the current methodological approach adopted by ANEEL and aligned with the principles of incentive regulation. An empirical study focusing on the residential class of electricity consumption demonstrated that it is possible to adopt a specific methodology for determining regulatory default targets and that the fuzzy inference approach can meet the necessary premises to ensure that the principles of incentive regulation and the establishment of regulatory targets are consistent with the reality of each electricity distribution utility. Full article
(This article belongs to the Section C: Energy Economics and Policy)
Show Figures

Figure 1

Figure 1
<p>Fuzzy inference system (Mamdani-type FIS).</p>
Full article ">Figure 2
<p>General view of the fuzzy inference-based model to establish regulatory targets for limiting or reducing electricity default rates in Brazil.</p>
Full article ">Figure 3
<p>Fuzzy groups created to classify each input and output variables.</p>
Full article ">Figure 4
<p>Fuzzy groups.</p>
Full article ">Figure 5
<p>Fuzzy rules for residential consumption class. Notation: VL—very low, L—low, M—moderate, H—high, VH—very high, and shaded cells—rules with weight 0.6.</p>
Full article ">Figure 6
<p>Choice of the defuzzification method: centroid.</p>
Full article ">Figure 7
<p>Comparison of ANEEL regulatory targets, default estimates by the fuzzy inference system (FIS), and actual default rates declared by distribution utilities.</p>
Full article ">Figure 8
<p>Example of applying fuzzy rules to a specific input.</p>
Full article ">Figure 9
<p>Comparison of actual default rates, ANEEL regulatory targets, and fuzzy inference default targets: 18 distribution utilities with the lowest residential default rates.</p>
Full article ">Figure 10
<p>Comparison of actual default rates, ANEEL regulatory targets, and fuzzy inference default targets: 21 distribution utilities with the moderate residential default rates.</p>
Full article ">Figure 11
<p>Comparison of actual default rates, ANEEL regulatory targets, and fuzzy inference default targets: 20 distribution utilities with highest residential default rates.</p>
Full article ">
19 pages, 7742 KiB  
Article
Cascaded Fuzzy Reward Mechanisms in Deep Reinforcement Learning for Comprehensive Path Planning in Textile Robotic Systems
by Di Zhao, Zhenyu Ding, Wenjie Li, Sen Zhao and Yuhong Du
Appl. Sci. 2024, 14(2), 851; https://doi.org/10.3390/app14020851 - 19 Jan 2024
Viewed by 1001
Abstract
With the rapid advancement of industrial automation and artificial intelligence technologies, particularly in the textile industry, robotic technology is increasingly challenged with intelligent path planning and executing high-precision tasks. This study focuses on the automatic path planning and yarn-spool-assembly tasks of textile robotic [...] Read more.
With the rapid advancement of industrial automation and artificial intelligence technologies, particularly in the textile industry, robotic technology is increasingly challenged with intelligent path planning and executing high-precision tasks. This study focuses on the automatic path planning and yarn-spool-assembly tasks of textile robotic arms, proposing an end-to-end planning and control model that integrates deep reinforcement learning. The innovation of this paper lies in the introduction of a cascaded fuzzy reward system, which is integrated into the end-to-end model to enhance learning efficiency and reduce ineffective exploration, thereby accelerating the convergence of the model. A series of experiments conducted in a simulated environment demonstrate the model’s exceptional performance in yarn-spool-assembly tasks. Compared to traditional reinforcement learning methods, our model shows potential advantages in improving task success rates and reducing collision rates. The cascaded fuzzy reward system, a core component of our end-to-end deep reinforcement learning model, offers a novel and more robust solution for the automated path planning of robotic arms. In summary, the method proposed in this study provides a new perspective and potential applications for industrial automation, especially in the operation of robotic arms in complex and uncertain environments. Full article
(This article belongs to the Topic Industrial Robotics: 2nd Volume)
Show Figures

Figure 1

Figure 1
<p>Example of a fully overlapping triangle affiliation function.</p>
Full article ">Figure 2
<p>Model-free end-to-end DRL framework.</p>
Full article ">Figure 3
<p>RGB-D images captured using Kinect in the simulation environment; (<b>a</b>,<b>c</b>,<b>e</b>) are RGB images and (<b>b</b>,<b>d</b>,<b>f</b>) are grayscale maps representing depths.</p>
Full article ">Figure 4
<p>Two-layer fuzzy system.</p>
Full article ">Figure 5
<p>Output of the fuzzy logic system during the initial stage: The figure on the left represents the first layer, while the right figure illustrates the second layer of the fuzzy logic system.</p>
Full article ">Figure 6
<p>Output of the fuzzy logic system during the mid-course obstacle-avoidance stage. The figure on the left depicts the first layer, and the right figure showcases the second layer of the fuzzy logic system.</p>
Full article ">Figure 7
<p>Output of the fuzzy logic system during the alignment and placement stage: The figure on the left shows the first layer, whereas the right figure displays the second layer of the fuzzy logic system.</p>
Full article ">Figure 8
<p>Continuous trajectory planning task for a textile robot.</p>
Full article ">Figure 9
<p>Performance comparison between end-to-end DDPG and baseline DDPG.</p>
Full article ">Figure 10
<p>Training of different reward systems.</p>
Full article ">Figure 11
<p>Collision rates for different reward systems.</p>
Full article ">Figure 12
<p>Robotic arm’s path-planning process in real environments.</p>
Full article ">
29 pages, 8163 KiB  
Article
A Necessity-Based Optimization Approach for Closed-Loop Logistics Considering Carbon Emission Penalties and Rewards under Uncertainty
by Botang Li, Kaiyuan Liu, Qiong Chen, Yui-yip Lau and Maxim A. Dulebenets
Mathematics 2023, 11(21), 4516; https://doi.org/10.3390/math11214516 - 1 Nov 2023
Cited by 7 | Viewed by 1224
Abstract
The recycling of waste products can bring enormous economic and environmental benefits to supply chain participants. Under the government’s reward and punishment system, the manufacturing industry is facing unfolded pressure to minimize carbon emissions. However, various factors related to the design of closed-loop [...] Read more.
The recycling of waste products can bring enormous economic and environmental benefits to supply chain participants. Under the government’s reward and punishment system, the manufacturing industry is facing unfolded pressure to minimize carbon emissions. However, various factors related to the design of closed-loop logistics networks are uncertain in nature, including demand, facility capacity, transportation cost per unit of product per kilometer, landfill cost, unit carbon penalty cost, and carbon reward amount. As such, this study proposes a new fuzzy programming model for closed-loop supply chain network design which directly relies on fuzzy methods based on the necessity measure. The objective of the proposed optimization model is to minimize the total cost of the network and the sum of carbon rewards and penalties when selecting facility locations and transportation routes between network nodes. Based on the characteristics of the problem, a genetic algorithm based on variant priority encoding is proposed as a solution. This new solution encoding method can make up for the shortcomings of the four traditional encoding methods (i.e., Prüfer number-based encoding, spanning tree-based encoding, forest data structure-based encoding, and priority-based encoding) to speed up the computational time of the solution algorithm. Several alternative solution approaches were considered to evaluate the proposed algorithm including the precision optimization method (CPLEX) and priority-based encoding genetic algorithm. The results of numerous experiments indicated that even for large-scale numerical examples, the proposed algorithm can create optimal and high-quality solutions within acceptable computational time. The applicability of the model was demonstrated through a sensitivity analysis which was conducted by changing the parameters of the model and providing some important management insights. When external parameters change, the solution of the model maintains a certain level of satisfaction conservatism. At the same time, the changes in the penalty cost and reward amount per unit of carbon emissions have a significant impact on the carbon penalty revenue and total cost. The results of this study are expected to provide scientific support to relevant supply chain enterprises and stakeholders. Full article
Show Figures

Figure 1

Figure 1
<p>Closed-loop logistics network structure.</p>
Full article ">Figure 2
<p>VPGA algorithm solution flow chart.</p>
Full article ">Figure 3
<p>The VPGA convergence patterns for test instance 1.</p>
Full article ">Figure 4
<p>Changes in total cost and carbon penalty/revenue (<span class="html-italic">pu</span> = <span class="html-italic">re</span> = 0.5).</p>
Full article ">Figure 5
<p>Logistics network structure.</p>
Full article ">Figure 6
<p>Changes in total cost and carbon penalty revenue (<math display="inline"><semantics> <mrow> <mi>p</mi> <mi>u</mi> <mo> </mo> </mrow> </semantics></math>=<math display="inline"><semantics> <mrow> <mo> </mo> <mi>r</mi> <mi>e</mi> </mrow> </semantics></math> = 1).</p>
Full article ">Figure 7
<p>Changes in total cost and carbon penalty revenue (<math display="inline"><semantics> <mrow> <mi>p</mi> <mi>u</mi> <mo> </mo> </mrow> </semantics></math>=<math display="inline"><semantics> <mrow> <mo> </mo> <mi>r</mi> <mi>e</mi> </mrow> </semantics></math> = 2).</p>
Full article ">Figure 8
<p>Changes in total cost and carbon penalty revenue (<math display="inline"><semantics> <mrow> <mi>p</mi> <mi>u</mi> <mo> </mo> </mrow> </semantics></math>=<math display="inline"><semantics> <mrow> <mo> </mo> <mi>r</mi> <mi>e</mi> </mrow> </semantics></math> = 3).</p>
Full article ">Figure 9
<p>Changes in total cost and carbon penalty revenue (<math display="inline"><semantics> <mrow> <mi>p</mi> <mi>u</mi> <mo> </mo> </mrow> </semantics></math>=<math display="inline"><semantics> <mrow> <mo> </mo> <mi>r</mi> <mi>e</mi> </mrow> </semantics></math> = 4).</p>
Full article ">Figure 10
<p>Changes in total cost and carbon penalty revenue (<math display="inline"><semantics> <mrow> <mi>p</mi> <mi>u</mi> </mrow> </semantics></math> = 2<math display="inline"><semantics> <mrow> <mo> </mo> <mi mathvariant="normal">a</mi> <mi mathvariant="normal">n</mi> <mi mathvariant="normal">d</mi> <mo> </mo> <mi>r</mi> <mi>e</mi> </mrow> </semantics></math> = 1).</p>
Full article ">Figure 11
<p>Changes in total cost and carbon penalty revenue (<math display="inline"><semantics> <mrow> <mi>p</mi> <mi>u</mi> </mrow> </semantics></math> = 3<math display="inline"><semantics> <mrow> <mo> </mo> <mi mathvariant="normal">a</mi> <mi mathvariant="normal">n</mi> <mi mathvariant="normal">d</mi> <mo> </mo> <mi>r</mi> <mi>e</mi> </mrow> </semantics></math> = 2).</p>
Full article ">Figure 12
<p>Changes in total cost and carbon penalty revenue (<math display="inline"><semantics> <mrow> <mi>p</mi> <mi>u</mi> </mrow> </semantics></math> = 4<math display="inline"><semantics> <mrow> <mo> </mo> <mi mathvariant="normal">a</mi> <mi mathvariant="normal">n</mi> <mi mathvariant="normal">d</mi> <mo> </mo> <mi>r</mi> <mi>e</mi> </mrow> </semantics></math> = 2).</p>
Full article ">Figure 13
<p>Changes in total cost and carbon penalty revenue (<math display="inline"><semantics> <mrow> <mi>p</mi> <mi>u</mi> </mrow> </semantics></math> = 4<math display="inline"><semantics> <mrow> <mo> </mo> <mi mathvariant="normal">a</mi> <mi mathvariant="normal">n</mi> <mi mathvariant="normal">d</mi> <mo> </mo> <mi>r</mi> <mi>e</mi> </mrow> </semantics></math> = 3.5).</p>
Full article ">
23 pages, 888 KiB  
Article
A Meta Reinforcement Learning Approach for SFC Placement in Dynamic IoT-MEC Networks
by Shuang Guo, Yarong Du and Liang Liu
Appl. Sci. 2023, 13(17), 9960; https://doi.org/10.3390/app13179960 - 3 Sep 2023
Cited by 1 | Viewed by 1448
Abstract
In order to achieve reliability, security, and scalability, the request flow in the Internet of Things (IoT) needs to pass through the service function chain (SFC), which is composed of series-ordered virtual network functions (VNFs), then reach the destination application in multiaccess edge [...] Read more.
In order to achieve reliability, security, and scalability, the request flow in the Internet of Things (IoT) needs to pass through the service function chain (SFC), which is composed of series-ordered virtual network functions (VNFs), then reach the destination application in multiaccess edge computing (MEC) for processing. Since there are usually multiple identical VNF instances in the network and the network environment of IoT changes dynamically, placing the SFC for the IoT request flow is a significant challenge. This paper decomposes the dynamic SFC placement problem of the IoT-MEC network into two subproblems: VNF placement and path determination of routing. We first formulate these two subproblems as Markov decision processes. We then propose a meta reinforcement learning and fuzzy logic-based dynamic SFC placement approach (MRLF-SFCP). The MRLF-SFCP contains an inner model that focuses on making SFC placement decisions and an outer model that focuses on learning the initial parameters considering the dynamic IoT-MEC environment. Specifically, the approach uses fuzzy logic to pre-evaluate the link status information of the network by jointly considering available bandwidth, delay, and packet loss rate, which is helpful for model training and convergence. In comparison to existing algorithms, simulation results demonstrate that the MRLF-SFCP algorithm exhibits superior performance in terms of traffic acceptance rate, throughput, and the average reward. Full article
(This article belongs to the Special Issue Advanced Artificial Intelligence Theories and Applications)
Show Figures

Figure 1

Figure 1
<p>A schematic diagram of IoT-MEC networks.</p>
Full article ">Figure 2
<p>A framework for placing dynamic SFCs based on meta reinforcement learning and fuzzy logic.</p>
Full article ">Figure 3
<p>The procedure of link evaluation.</p>
Full article ">Figure 4
<p>The general training process of outer model.</p>
Full article ">Figure 5
<p>Dynamic SFC placement framework based on DDQN.</p>
Full article ">Figure 6
<p>The acceptance rate in different network topologies.</p>
Full article ">Figure 7
<p>The average reward of IoT-SRs in different network topologies.</p>
Full article ">Figure 8
<p>The acceptance rate of IoT-SRs of the RN.</p>
Full article ">Figure 9
<p>The mean reward of IoT-SRs within the RN.</p>
Full article ">Figure 10
<p>The acceptance rate of IoT-SRs in the SWN.</p>
Full article ">Figure 11
<p>Average reward of IoT-SRs in the SWN.</p>
Full article ">Figure 12
<p>Success acceptance rate of IoT-SRs in the SFN.</p>
Full article ">Figure 13
<p>Average reward of IoT-SRs in the SFN.</p>
Full article ">
19 pages, 1868 KiB  
Article
Unleashing Intrapreneurial Behavior: Exploring Configurations of Influencing Factors among Grassroots Employees
by Di Ye, Wenlong Xie and Linlin Zheng
Behav. Sci. 2023, 13(9), 724; https://doi.org/10.3390/bs13090724 - 30 Aug 2023
Cited by 2 | Viewed by 1570
Abstract
Effectively promoting employees’ intrapreneurial behavior has become the focus of enterprises. This study takes the middle and grassroots employees in enterprises as subjects and explores the configuration effect of multiple influencing factors on employees’ intrapreneurial behavior. Based on employee expectation theory and individual-environment [...] Read more.
Effectively promoting employees’ intrapreneurial behavior has become the focus of enterprises. This study takes the middle and grassroots employees in enterprises as subjects and explores the configuration effect of multiple influencing factors on employees’ intrapreneurial behavior. Based on employee expectation theory and individual-environment matching theory, this study collates six influencing factors: entrepreneurial self-efficacy, entrepreneurial competence, task school level, perceived value, management support, and reward mechanism. A total of 163 samples were obtained, and the qualitative comparative analysis method based on fuzzy set was used to analyze the influence mechanism and result path of employees’ intrapreneurial behavior from the perspective of the interaction between individual factors and organizational factors. Six influencing paths of employees’ high intrapreneurial behavior were found, which can be divided into ability-driven and value-driven factors, revealing that the six factors can produce equivalent results in different configurations. Furthermore, five influencing paths of employees’ non-high intrapreneurial behavior were divided into three types: ability obstacle type, perception obstacle type, and value obstacle type. These have an asymmetric causal relationship with employees’ high intrapreneurial behavior. This study provides management support for effectively stimulating employees’ intrapreneurial behavior. Full article
(This article belongs to the Section Organizational Behaviors)
Show Figures

Figure 1

Figure 1
<p>Research Model: Model of Drivers Influencing Employees’ Intrapreneurial Behavior.</p>
Full article ">
19 pages, 743 KiB  
Article
Analysing the Barriers Involved in Recycling the Textile Waste in India Using Fuzzy DEMATEL
by S. G. Ponnambalam, Bathrinath Sankaranarayanan, Koppiahraj Karuppiah, Shakthi Thinakaran, Pranesh Chandravelu and Hon Loong Lam
Sustainability 2023, 15(11), 8864; https://doi.org/10.3390/su15118864 - 31 May 2023
Cited by 8 | Viewed by 5272
Abstract
Post-consumer wastes from the textile industry are generally landfilled or incinerated. The dumping of large amounts of textile waste has resulted in severe environmental problems. Advancements in technologies have called for textile recycling; however, the level of embracement made by the textile industry [...] Read more.
Post-consumer wastes from the textile industry are generally landfilled or incinerated. The dumping of large amounts of textile waste has resulted in severe environmental problems. Advancements in technologies have called for textile recycling; however, the level of embracement made by the textile industry towards textile recycling is hampered by myriad factors. The scope of this study lies in identifying and analyzing multiple barriers to implementing textile recycling in India, encompassing all subsets of sustainability, i.e., social, economic, and environmental. The barriers are then evaluated using a Multiple Criteria Decision Making (MCDM) approach to identify the significant barriers. A trapezoidal fuzzy-DEMATEL methodology was executed to not only find the most influential barriers but also to find the cause-effect nature between every barrier. The outcome of the study indicates a lack of successful recycling business models, poor demand for recycled textiles goods, recycled products may not replace new products, lack of support for waste management in the industry, and absence of tax relief and rewarding policies as the top five barriers to textile waste recycling. This insight could help influence the decision of future policymakers in the field. Another aspect of the issue of pollution in the textile industry is the recent trend of fast fashion and the enormous amount of waste produced by overconsumption. The Sustainability Development Goal (SDG) 12 which is to ensure responsible production and consumption plays a key role in this sector. Full article
(This article belongs to the Section Sustainable Products and Services)
Show Figures

Figure 1

Figure 1
<p>The methodology adopted.</p>
Full article ">Figure 2
<p>Cause-effect Diagram.</p>
Full article ">
32 pages, 4255 KiB  
Review
Reviewing Evolution of Learning Functions and Semantic Information Measures for Understanding Deep Learning
by Chenguang Lu
Entropy 2023, 25(5), 802; https://doi.org/10.3390/e25050802 - 15 May 2023
Cited by 2 | Viewed by 1883
Abstract
A new trend in deep learning, represented by Mutual Information Neural Estimation (MINE) and Information Noise Contrast Estimation (InfoNCE), is emerging. In this trend, similarity functions and Estimated Mutual Information (EMI) are used as learning and objective functions. Coincidentally, EMI is essentially the [...] Read more.
A new trend in deep learning, represented by Mutual Information Neural Estimation (MINE) and Information Noise Contrast Estimation (InfoNCE), is emerging. In this trend, similarity functions and Estimated Mutual Information (EMI) are used as learning and objective functions. Coincidentally, EMI is essentially the same as Semantic Mutual Information (SeMI) proposed by the author 30 years ago. This paper first reviews the evolutionary histories of semantic information measures and learning functions. Then, it briefly introduces the author’s semantic information G theory with the rate-fidelity function R(G) (G denotes SeMI, and R(G) extends R(D)) and its applications to multi-label learning, the maximum Mutual Information (MI) classification, and mixture models. Then it discusses how we should understand the relationship between SeMI and Shannon’s MI, two generalized entropies (fuzzy entropy and coverage entropy), Autoencoders, Gibbs distributions, and partition functions from the perspective of the R(G) function or the G theory. An important conclusion is that mixture models and Restricted Boltzmann Machines converge because SeMI is maximized, and Shannon’s MI is minimized, making information efficiency G/R close to 1. A potential opportunity is to simplify deep learning by using Gaussian channel mixture models for pre-training deep neural networks’ latent layers without considering gradients. It also discusses how the SeMI measure is used as the reward function (reflecting purposiveness) for reinforcement learning. The G theory helps interpret deep learning but is far from enough. Combining semantic information theory and deep learning will accelerate their development. Full article
(This article belongs to the Special Issue Entropy: The Cornerstone of Machine Learning)
Show Figures

Figure 1

Figure 1
<p>The distinctions and relations between four types of learning [<a href="#B31-entropy-25-00802" class="html-bibr">31</a>,<a href="#B34-entropy-25-00802" class="html-bibr">34</a>].</p>
Full article ">Figure 2
<p>Illustrating a GPS device’s positioning with a deviation. We predict the probability distribution of <span class="html-italic">x</span> according to <span class="html-italic">y<sub>j</sub></span> and the prior knowledge <span class="html-italic">P</span>(<span class="html-italic">x</span>). The red star represents the most probable position.</p>
Full article ">Figure 3
<p>The semantic information conveyed by <span class="html-italic">y<sub>j</sub></span> about <span class="html-italic">x<sub>i</sub></span> decreases with the deviation or distortion increasing. The larger the deviation is, the less information there is.</p>
Full article ">Figure 4
<p>The information rate-fidelity function <span class="html-italic">R</span>(<span class="html-italic">G</span>) for binary communication. Any <span class="html-italic">R</span>(<span class="html-italic">G</span>) function is a bowl-like function. There is a point at which <span class="html-italic">R</span>(<span class="html-italic">G</span>) = <span class="html-italic">G</span> (<span class="html-italic">s</span> = 1). For given <span class="html-italic">R</span>, two anti-functions exist: <span class="html-italic">G</span><sup>-</sup>(<span class="html-italic">R</span>) and <span class="html-italic">G</span><sup>+</sup>(<span class="html-italic">R</span>).</p>
Full article ">Figure 5
<p>Illustrating the medical test and the signal detection. We choose <span class="html-italic">y<sub>j</sub></span> according to <span class="html-italic">z</span> ∈ <span class="html-italic">C<sub>j</sub></span>. The task is to find the dividing point <span class="html-italic">z</span>’ that results in MaxMI between <span class="html-italic">X</span> and <span class="html-italic">Y</span>.</p>
Full article ">Figure 6
<p>The MMI classification with a very bad initial partition. The convergence is very fast and stable without considering gradients. (<b>a</b>) The very bad initial partition. (<b>b</b>) The partition after the first iteration. (<b>c</b>) The partition after the second iteration. (<b>d</b>) The mutual information changes with iterations.</p>
Full article ">Figure 7
<p>Comparing EM and E3M algorithms with an example that is hard to converge. The EM algorithm needs about 340 iterations, whereas the E3M algorithm needs about 240 iterations. In the convergent process, complete data log-likelihood <span class="html-italic">Q</span> is not monotonously increasing. <span class="html-italic">H</span>(<span class="html-italic">P||P<sub>θ</sub></span>) decreases with <span class="html-italic">R − G.</span> (<b>a</b>) Initial components with (<span class="html-italic">µ</span><sub>1</sub>, <span class="html-italic">µ</span><sub>2</sub>) = (80, 95). (<b>b</b>) Globally convergent two components. (<b>c</b>) <span class="html-italic">Q</span>, <span class="html-italic">R</span>, <span class="html-italic">G</span>, and <span class="html-italic">H</span>(<span class="html-italic">P||P<sub>θ</sub></span>) changes with iterations (initialization: (<span class="html-italic">µ</span><sub>1</sub>, <span class="html-italic">µ</span><sub>2</sub>, <span class="html-italic">σ</span><sub>1</sub>, <span class="html-italic">σ</span><sub>2</sub>, <span class="html-italic">P</span>(<span class="html-italic">y</span><sub>1</sub>)) = (80, 95, 5, 5, 0.5)).</p>
Full article ">Figure 7 Cont.
<p>Comparing EM and E3M algorithms with an example that is hard to converge. The EM algorithm needs about 340 iterations, whereas the E3M algorithm needs about 240 iterations. In the convergent process, complete data log-likelihood <span class="html-italic">Q</span> is not monotonously increasing. <span class="html-italic">H</span>(<span class="html-italic">P||P<sub>θ</sub></span>) decreases with <span class="html-italic">R − G.</span> (<b>a</b>) Initial components with (<span class="html-italic">µ</span><sub>1</sub>, <span class="html-italic">µ</span><sub>2</sub>) = (80, 95). (<b>b</b>) Globally convergent two components. (<b>c</b>) <span class="html-italic">Q</span>, <span class="html-italic">R</span>, <span class="html-italic">G</span>, and <span class="html-italic">H</span>(<span class="html-italic">P||P<sub>θ</sub></span>) changes with iterations (initialization: (<span class="html-italic">µ</span><sub>1</sub>, <span class="html-italic">µ</span><sub>2</sub>, <span class="html-italic">σ</span><sub>1</sub>, <span class="html-italic">σ</span><sub>2</sub>, <span class="html-italic">P</span>(<span class="html-italic">y</span><sub>1</sub>)) = (80, 95, 5, 5, 0.5)).</p>
Full article ">Figure 8
<p>Comparing a typical neuron and a neuron in a CMMM. (<b>a</b>) A typical neuron in neural networks. (<b>b</b>) A neuron in the CMMM and its optimization.</p>
Full article ">Figure 9
<p>Illustrating population death age control for measuring purposive information. <span class="html-italic">P(x|a<sub>j</sub></span>) approximates to <span class="html-italic">P(x|θ<sub>j</sub></span>) = <span class="html-italic">P(x|θ<sub>j</sub></span>, <span class="html-italic">s</span> = 1) for information efficiency <span class="html-italic">G/R</span> = 1. <span class="html-italic">G</span> and <span class="html-italic">R</span> are close to their maxima as <span class="html-italic">P(x|a<sub>j</sub></span>) approximates to <span class="html-italic">P(x|θ<sub>j</sub></span>, <span class="html-italic">s</span> = 20).</p>
Full article ">
17 pages, 807 KiB  
Article
Dynamic Evaluation of Energy Carbon Efficiency in the Logistics Industry Based on Catastrophe Progression
by Xiaohong Yin, Yufei Wu and Qiang Liu
Sustainability 2023, 15(6), 5574; https://doi.org/10.3390/su15065574 - 22 Mar 2023
Cited by 2 | Viewed by 1334
Abstract
The logistics industry has an irreplaceable role in promoting Chinese economic development, and its carbon emissions have become a hot topic of academic research. However, more research needs to be conducted on this. This study is based on establishing an evaluation index system [...] Read more.
The logistics industry has an irreplaceable role in promoting Chinese economic development, and its carbon emissions have become a hot topic of academic research. However, more research needs to be conducted on this. This study is based on establishing an evaluation index system for the efficiency of energy carbon emissions in the Chinese logistics industry. The catastrophe progression method was used to evaluate this statically. A dynamic evaluation model was also established based on the characteristics of fuzzy rewards and punishments. The results showed that the static values in the southeastern provinces of China were always between 0.9 and 1, and there was a significant increase in the dynamic values under the fuzzy reward and punishment scenario. Provinces in the southwest fluctuated between 0.8 and 0.95, while the dynamic values did not increase much. In the northern provinces, the static assessment values were consistently between 0.7 and 0.9, while the dynamic values were decreasing. It is therefore important to reward provinces with high static assessment values and penalize those with low static assessment values. The perspective of the characteristics of fuzzy rewards and punishments is also essential for fair and equitable management, reward and punishment in the different provinces in the study. Full article
(This article belongs to the Special Issue Circular Economy and Logistics)
Show Figures

Figure 1

Figure 1
<p>Diagram of indicator calculation research process.</p>
Full article ">Figure 2
<p>Research flow chart.</p>
Full article ">Figure 3
<p>Comparison chart of methods.</p>
Full article ">
20 pages, 1499 KiB  
Article
Human–Computer Interaction and Participation in Software Crowdsourcing
by Habib Ullah Khan, Farhad Ali, Yazeed Yasin Ghadi, Shah Nazir, Inam Ullah and Heba G. Mohamed
Electronics 2023, 12(4), 934; https://doi.org/10.3390/electronics12040934 - 13 Feb 2023
Cited by 3 | Viewed by 2932
Abstract
Improvements in communication and networking technologies have transformed people’s lives and organizations’ activities. Web 2.0 innovation has provided a variety of hybridized applications and tools that have changed enterprises’ functional and communication processes. People use numerous platforms to broaden their social contacts, select [...] Read more.
Improvements in communication and networking technologies have transformed people’s lives and organizations’ activities. Web 2.0 innovation has provided a variety of hybridized applications and tools that have changed enterprises’ functional and communication processes. People use numerous platforms to broaden their social contacts, select items, execute duties, and learn new things. Context: Crowdsourcing is an internet-enabled problem-solving strategy that utilizes human–computer interaction to leverage the expertise of people to achieve business goals. In crowdsourcing approaches, three main entities work in collaboration to solve various problems. These entities are requestors (job providers), platforms, and online users. Tasks are announced by requestors on crowdsourcing platforms, and online users, after passing initial screening, are allowed to work on these tasks. Crowds participate to achieve various rewards. Motivation: Crowdsourcing is gaining importance as an alternate outsourcing approach in the software engineering industry. Crowdsourcing application development involves complicated tasks that vary considerably from the micro-tasks available on platforms such as Amazon Mechanical Turk. To obtain the tangible opportunities of crowdsourcing in the realm of software development, corporations should first grasp how this technique works, what problems occur, and what factors might influence community involvement and co-creation. Online communities have become more popular recently with the rise in crowdsourcing platforms. These communities concentrate on specific problems and help people with solving and managing these problems. Objectives: We set three main goals to research crowd interaction: (1) find the appropriate characteristics of social crowd utilized for effective software crowdsourcing, (2) highlight the motivation of a crowd for virtual tasks, and (3) evaluate primary participation reasons by assessing various crowds using Fuzzy AHP and TOPSIS method. Conclusion: We developed a decision support system to examine the appropriate reasons of crowd participation in crowdsourcing. Rewards and employments were evaluated as the primary motives of crowds for accomplishing tasks on crowdsourcing platforms, knowledge sharing was evaluated as the third reason, ranking was the fourth, competency was the fifth, socialization was sixth, and source of inspiration was the seventh. Full article
(This article belongs to the Section Networks)
Show Figures

Figure 1

Figure 1
<p>Crowdsourcing entities and its task assignment strategy representation.</p>
Full article ">Figure 2
<p>Average and normalized weights of criteria.</p>
Full article ">Figure 3
<p>Ranking and performance of alternatives.</p>
Full article ">
25 pages, 2352 KiB  
Article
A New Fuzzy Reinforcement Learning Method for Effective Chemotherapy
by Fawaz E. Alsaadi, Amirreza Yasami, Christos Volos, Stelios Bekiros and Hadi Jahanshahi
Mathematics 2023, 11(2), 477; https://doi.org/10.3390/math11020477 - 16 Jan 2023
Cited by 3 | Viewed by 2023
Abstract
A key challenge for drug dosing schedules is the ability to learn an optimal control policy even when there is a paucity of accurate information about the systems. Artificial intelligence has great potential for shaping a smart control policy for the dosage of [...] Read more.
A key challenge for drug dosing schedules is the ability to learn an optimal control policy even when there is a paucity of accurate information about the systems. Artificial intelligence has great potential for shaping a smart control policy for the dosage of drugs for any treatment. Motivated by this issue, in the present research paper a Caputo–Fabrizio fractional-order model of cancer chemotherapy treatment was elaborated and analyzed. A fix-point theorem and an iterative method were implemented to prove the existence and uniqueness of the solutions of the proposed model. Afterward, in order to control cancer through chemotherapy treatment, a fuzzy-reinforcement learning-based control method that uses the State-Action-Reward-State-Action (SARSA) algorithm was proposed. Finally, so as to assess the performance of the proposed control method, the simulations were conducted for young and elderly patients and for ten simulated patients with different parameters. Then, the results of the proposed control method were compared with Watkins’s Q-learning control method for cancer chemotherapy drug dosing. The results of the simulations demonstrate the superiority of the proposed control method in terms of mean squared error, mean variance of the error, and the mean squared of the control action—in other words, in terms of the eradication of tumor cells, keeping normal cells, and the amount of usage of the drug during chemotherapy treatment. Full article
Show Figures

Figure 1

Figure 1
<p>Sensitivity analysis.</p>
Full article ">Figure 2
<p>Structure of reinforcement learning.</p>
Full article ">Figure 3
<p>Membership functions used for fuzzy-sets. (<b>a</b>) Trapezoidal-shape membership function. (<b>b</b>) Z-shape membership function for upper bound. (<b>c</b>) Z-shape membership function for lower bound.</p>
Full article ">Figure 4
<p>The result of simulation using the proposed control method for young patient without uncertainty. (<b>a</b>) Normal-cells (<b>b</b>) Tumor-cells (<b>c</b>) Immune-cells (<b>d</b>) Concentration of cells (<b>e</b>) Control action.</p>
Full article ">Figure 5
<p>The result of simulation using the proposed control method for young patient with uncertainty. (<b>a</b>) Normal-cells (<b>b</b>) Tumor-cells (<b>c</b>) Immune-cells (<b>d</b>) Concentration of cells (<b>e</b>) Control action.</p>
Full article ">Figure 6
<p>The result of simulation using the proposed control method for elderly patient without uncertainty. (<b>a</b>) Normal-cells (<b>b</b>) Tumor-cells (<b>c</b>) Immune-cells (<b>d</b>) Concentration of cells (<b>e</b>) Control action.</p>
Full article ">Figure 7
<p>The result of simulation using the proposed control method for elderly patient with uncertainty. (<b>a</b>) Normal-cells (<b>b</b>) Tumor-cells (<b>c</b>) Immune-cells (<b>d</b>) Concentration of cells (<b>e</b>) Control action.</p>
Full article ">
Back to TopTop