Виноградов Г.П. (wgp272ng@mail.ru) - Тверской государственный технический университет (профессор), Тверь, Россия, доктор технических наук | |
Ключевые слова: многоагентная система, агент, модель принятия решений, ситуация выбора, моделирование |
|
Keywords: multiagents systems, agent, decision-making model, choice situation, modeling |
|
|
Human behavior (hereinafter purposeful agent – CA) is the result of his cognitive activity. In psychology, cognitive activity is regarded as forming a mental image (the result of perception) and models of knowable and analyzed objects, processes, phenomena. According to generated model and based on fixed mind installations, rules, customs and past life experience, knowledge, the subject produces his own attitude to the observed and analyzed. Then the subject forms his behavior. It depends on this attitude and adequacy degree evaluation between the object and the result of mental visualization and analysis. He also develops own action program (he can vary degree of his participation up to the complete non-participation) in the events, processes, phenomena, which he considers important for himself. Therefore, one of the approaches to overcome the shortcomings of traditional methods of describing the agents’ behavior in organizational systems is the modification of rational behavior models. It is achieved by including the so-called unobserved variables. They are characteristics of human behavior used in psychology, sociology, etc. For example, we introduce the r parameter. It will characterize the type of an agent, and (r otherwise can be called the agent characteristic) [1]. The objective function of agent will depend on the environment state , performance , that achieved by implementation of the agent’s action cÎC, management of the center u(·)ÎU and r type agent. If we follow the hypothesis of rational behavior of an agent, then the agent will try to get a result that will maximize his objective function. This allows us to speculate about the function with information: This approach must use some idealized design (model) that describes the behavior of CA in decision-making. Purposeful behavior model Purposeful behavior is always associated with the selection that occurs in a situation of purposeful state [2–4]. Purposeful state consists of the following components. · Entity that makes a choice (purposeful agent – CA). · Selection setting (S), which is represented by the set of elements and their essential properties. Any of property changes may cause or produce a change in the purposeful choice state. · Available action methods of k-th agent, which he owns and can use to achieve the i-th result (also called alternatives). · Possible surrounded by S scores (significant for CA) – . · The method of evaluating properties of the obtained results when choosing the mode of action. · Restrictions that reflect the requirements to the output variables and control actions, imposed by the purposeful choice situation. · Subjective decision-making situation model. It is represented as a set of relations describing the dependence of control actions, parameters and disturbances with the output variables. For described components we introduce measures that will be used to assess the purposeful state. 1. We assume that the CA is able to allocate factors. They are environment characteristics . The CA evaluates influence of each factor using linguistic variable “factor influence degree”. CA formalizes the ideas about the influence of selected factors on the results . We assume that for this purpose the CA uses an approximation in the form of production rules, which are as follows: If is and if is ... and if is , then , (1) where R is a number of production rules, r is a number of current production rule, is a clear function. It reflects the CA’s view of functional relationship of input factors with possible results for the r-th rule; is fuzzy variables defined on . 2. The modes of action that are known (available) by CA. is a function of the environment state parameters taken into consideration by the CA, functional and morphological properties of the system. Then a set of assumptions about their possible values forms a scenario of a possible environment state, system functionality and possibility of morphological changes in the system. For example, the implementation of scenarios with rules (1) allows forming an idea of the possible outcome. Hence, it determines the nonequivalence when choosing the mode of action that can be described as the confidence degree of the necessity of mode of action. This estimate can be described by linguistic variables. This measure is an individual characteristic of the CA. It can be changed as a result of learning and gaining experience. 3. Possible results for a given choice environment. When the agent makes a decision in a purposeful state situation to achieve a result, the choice of mode of action is connected with quantitative assessment of selected decision properties as shown in [5]. Selecting a list of properties and their parameters depends on the CA. We represent the possible outcomes for a given choice environment of CA as , where is a set of possible outcomes when choosing the j-th mode of action, is a set of results taken into account by k-th CA. 4. Value of the results. The CA is able to compare the benefits that he gets when receiving different types of stimulation results with labor costs to achieve this. This assumption determines the existence of the indicator. The value of j-th result type can be evaluated using the following linguistic variable . 5. Effectiveness in the context of the result. It is a linguistic variable. It expresses the CA’s individual assessment of the consequences of the choice regarding costs: financial, material, labor, labor intensity. is a degree of certainty, that some mode of action will lead to a result in S environment, if the CA will choose it. . It allows evaluating the effectiveness of the method chosen by the CA. 6. We define the value of a purposeful state of i-th result for k-th CA according to Sugeno rule [6]: . We can similarly estimate the value of a purposeful state for k-th CA according to the efficiency for i-th type of results: . 7. CA’s evaluation of purposeful state desirability using i-th result and its achieving efficiency in a choice situation is defined as a linguistic variable. The basis of the interaction between individuals and organizations is psychological and economic contracts. They reflect significant expectations of a person and relevant expectations of the organization. Consequently, we can define the following constraints: and , where and are expectations of the CA from the organization, which reflect the balance between costs and rewards for achieved results. When a reward function is fixed, according to the rational behavior hypothesis the CA generates a decision as where is an assessment of maximum win, which the CA could obtain by performing other work. If , we should expect that CA will choose ci=0. These rules mean that there is a CA that wants to get some result. It has several alternative ways to achieve the result with different efficiency. And its confidence in getting the desired result is significant. Fuzzy description model for agent’s purposeful behavior A goal is accomplished by tasks. And tasks are accomplished by outcomes [6, 7]. An outcome defines a desired result for the agent in a short period of time in certain s choice environment:
The task is the last desired result for the agent in the sequence of results With It means that the task is described as following: The goal is an unattainable result W. But it can be infinitely close in the sequence of tasks in time interval Moreover and the reasonableness of achieving the goal is (the membership function "reasonableness"). It approaches as n increases to 1 with no limit. This follows from progressivity properties of the purposeful intelligent system [2, 7]. Advancing the task, the expected specific value regarding the result increases monotonically in the time interval . And advancing the goal at the time interval :
Advancing the goal and the task, the majority of available actions for an agent and the set of possible outcomes can be fixed. When there is a need to create new ways of actions and new results in addition to old, their online search is organized. Then (2) Model of agent’s submissions about the purposeful choice state A purposeful choice situation can include: M is constraints, C is modes of action, i is normative or ideal elements, ie is a symbolic expressions of normative or ideal elements, W is an external environment. For the agent a purposeful choice situation exists in the form of representations. The agent forms a representation using following components: – available ways of acting; – possible outcomes from this acting; – possible states of the choice environment (possible values of uncontrolled variables that may affect the results of using modes of action, including representations of other agents); – probabilities that each possible state of choice environment is true; – the efficiency of each available mode of action for each possible outcome in every possible state of choice environment; – specific value of each possible outcome. There is a special aspect in forming representations by an agent. He uses not only the results of external environment monitoring, but also the results of monitoring the actions and results of other agents as an information for adjusting his perceptions about uncertain parameters. He tries to "explain" why they chose those actions. Let qÎW is a nature state. The agent with the number kÎK has interval information about this nature state. It is . The result of the system o=G(q, c) depends on the vector c=(c1, c2, ..., ck)ÎC’= of other agents’ actions associated with the k-th agent, ciÎCi, and q are nature states. We assume that the specific value of each agent’s choice depends on q nature state and o result of the system: We consider that K set of agents, their results {Ok (•)}, {Ck} set of feasible action, W set of possible values of nature states and G(•) function is a common knowledge. Each agent observes the action vector of all agents, the overall result and the results of all agents. The overall result and the specific value of each agent’s results depends on the actions of all agents (2). That is why we can use the game theory apparatus to describe their behavior. We denote the set of parametric Nash equilibrium (a parameter is the nature status value) as: If a set W0ÌΩ of possible nature status values is common knowledge among agents, then we obtain the following set of game equilibrium (assuming that they eliminate the uncertainty by calculating the maximum guaranteed result): We denote q(c)ÍW as a set of nature states, when the vector of agents’ modes of action is an equilibrium: . We denote as the vector of result values for agents, they consider it satisfying. When the specific values of the agents’ choice results g are the maximum (along with the observed o result), the set of nature status values is the following: . vector is determined by EVk for each kÎK. Then vector determines the required level of agents’ representation in a purposeful choice situation. According to [7] we assume that the agent is subjectively rational: 1) his interests are expressed as subjective value assessments of expected results using modes of action based on W purposeful choice situation representation; 2) agent’s behavior rationality involves striving to maximize the specific value of expected results. The agent chooses the mode of action under uncertainty and when the information is incomplete. While analyzing, the agent forms his own view of the choice situation as a hypothesis. He also considers the significance degree for the observed parameters and classifies them as for or against evidence. Then the agent searches for data to support the hypothesis and deny it. However, he uses only data that is necessary and sufficient for understanding the processes in his subject domain. If the facts refute the hypothesis, the agent modifies or revises it and includes the positive aspects from the old hypothesis. Data analysis allows generating questions (queries) to confirm the assumptions, information search for the answer. It is the basis for acceptance or rejection of the original submission. This strategy allows forming a consistent relationship between the observed parameters and agent’s representations when the source data is incomplete and unreliable. Definition 1. The agent is confident in his representation adequacy in the Ω type choice situation for G goal. And based on this he considers the choice of C mode of action helpful to achieve G goal. In this case: 1) the agent perceives a part of X situation characteristics for Ω choice; 2) for the other part of X situation characteristics he makes assumptions and shows the intention to prove (verify) their credibility; 3) in similar Ω type choice situations the agent always achieved G, when perceiving the presence (absence) X and aiming for G choosing C; 4) when the agent observed the absence (presence) of X in Ω choice situation, he never chosen C to achieve G goal based on this submission. The assumption is a default value of the observed characteristics or a description of a causal effect between the observable characteristics. Agent’s representations are characterized by conviction level. Definition 2. The agent is sure of his perceptions of Ω type choice situation for G goal. His conviction level is determined by the achievement frequency of G goal when choosing C mode of action based on his perceptions. Assessing the level of conviction varies from zero to one. If the number of unsuccessful attempts to achieve the G goal with C mode of action based on agent’s representations increases, then the level of conviction decreases (and vice versa). It becomes a trigger for the agent to make efforts for modifying representations or their complete reconstruction. It is a result of increasing doubts about the plausibility of made assumptions. The desire to verify the correctness of the assumptions is a measure of agent’s doubt. Definition 3. The agent makes efforts to prove (disprove) an assumption. These efforts define the extent of his doubts about the submissions of Ω type choice situation when aiming to the G goal. The definition 2 says that the level of conviction depends on the number of confirmations of the correct choice based on the submissions. According to theory of behavioral psychology, if this level increases, then the agent’s desire of verifying falls. The agent sees no sense in it. On the other hand, increase of the agent’s degree of doubt is a trigger to find additional arguments (counterarguments). The degree of conviction is a parameter that takes into account these two characteristics: Su a+b=1, where Su is a degree of conviction, Uu is a level of conviction (past experience), Ss is a degree of the agent’s doubt about the correctness of his perceptions of the choice situation, a and b are coefficients of significance for the agent’s experience and the necessity of evidence search. Assumption 1. When forming their views, agents use the argument apparatus to create a sequence of hypotheses converging to subjectively true [8]. Assumption 2. When considering the A event that is occurred, the agent believes that the B event is a cause. On the other hand, observing the A event, the agent makes a conclusion about the possibility of the B event. At the k-th agent has the following data sources about the state of the environment [9]: 1) prior private information (this type of information is called knowledge, experience); 2) the actions of other agents. Observing and assuming that the opponents act rationally, the agent can (with assumption of common knowledge) to asses q(c) information about nature state. Based on this information the choice of mode of action was made by the i-th agent (i≠k); 3) the g set of agents’ choice results. This information helps agents to make a conclusion about nature states when the observed result leads to the observed winnings; 4) the set rÍW of nature states, when the observed vector of agents’ actions leads to this observed o value of the system result: s(c, o)={qÎWêG(q, c)=o}. According to the assumption made above, the information in 2–4 entries is a common knowledge for agents. Then the following relation is fair: I(c, o, g)=q(c)Çs(x, o)Çh(g, o)ÍW. Based on the common knowledge and their own views each k-th agent can calculate the estimate of nature state value as the intersection of the I(c, o, g) common knowledge with its private information : [10]. We assume that is a mode of action of k-th agent to reach a result on step t, ct is a set of vectors of all agents’ actions in t steps. At the moment of reaching a result on step t the common knowledge among agents is: I(ct, ot, gt)=q(ct)Çs(ct, ot)Çh(gt, ot)ÍW. Based on all sources of information k-th agent in t periods can calculate the estimate of value of the nature state as the intersection of the I(ct, ot, gt) common knowledge, with its private information that corresponds with the previous period: Due to the rise of the specific value of the result for the agent when achieving the goal and according to the agent’s representations formation properties , where is evaluation of agent’s conviction in the adequacy of his representation structure. Thus, when the hypothesis of rational behavior and the hypothesis of common knowledge is true, there will be a convergent process of forming representations of the agent. Evaluation of awareness for the fuzzy choice problem When information and knowledge is incomplete, or there is a lack of time, the agent creates a model of the subject domain temporarily based on plausible assumptions. He does not have sufficient evidence for these assumptions (e.g., support information) [8, 9, 11, 12]. A set of such assumptions, ideas, views is for explaining the phenomena, processes and relationships between them in a particular subject domain. And they form a hypothetical concept of an agent, which is understood as a speculative, subjective knowledge. The conclusions (opinions) based on hypothetical concept (subjective theory) determine a conviction state of a decision maker. He is convinced in his subjective understanding of purposeful state. Conviction stability depends on the frequency and duration of successful results for the goal (O) due to the truth of accepted assumptions, hypotheses, the correctness of the rules of output construction (i.e., a hypothetical concept in general). If the result of solution based on the subjective perceptions of the decision maker does not match his expectations and creates a state of doubt about the purposeful state, he implements the non-monotonic conviction revision process. This process involves removing a error assumption and/or introducing a new assumption, a correction of output rules. Thus, new convictions are a result of new data gathered from the output system and the monitoring solutions system. Definition 4. We assume that a fuzzy set AÍX and BÍX, where X is a clear set. For each fuzzy set we define a-level sets: , , where and are functions of belonging. Their values reflect the agent's degree of confidence that an x element belongs A and B. Then a alternative will be more preferable than b alternative, if and only if . A is greater than B on a level. We denote a as a minimum value of a when the inequality is true. Then 1–a is the degree of confidence that a is more preferable than b and indifference when choosing a or b. By analogy, if is contained in (), then we say that A is contained in B on a level. If the r=1–a value increases (or a decreases), then the statement that A is greater than B increases (decreases). If a=0, any element belonging to a fuzzy set reliably belongs only to this set. Introduction a degree of conviction when comparing alternatives allows: – determining the degree of information sufficiency for decision-making. If the value of the degree of confidence is below a certain level, then the decision-making is delayed to gather more information; – determining the value of collected additional information for CA. It can be zero if the confidence level does not change after receiving. If the value r=1–a increased, then the information contributed to the increase of the representation degree of the CA about the choice situation. If , then either there is a misinformation, or the obtained data destroys the representation of the CA about the choice situation and there is a need in new data. Value allows determining the direction of information search. We suppose there are two statements: r@X is G and q@X is F, where F and G are predicates represented as fuzzy sets. Then, if GÌF, pÞq (p implies q), this means that the first statement is more informative than the second. Therefore, changing the awareness of CA leads to changing of his views, and to changing and supp A. And they can be used as an awareness measure of CA. Achieving GÌF effect requires increasing the number of the properties taken into account when describing. In addition, each additional property should increase the confidence level in distinguishing objects. Increasing the number of properties can lead to two situations. 1. The new information increases the confidence degree in GÌF. Id est the statement with the new property is more informative than the same statement without it. 2. When comparing two objects with the same number of evaluated properties, and another property is added, but its value for both objects is difficult to distinguish, then the extra information does not increase the degree of confidence in the legibility of objects, but does not reduce it. The article describes an approach to modeling decision-making by a person based on his subjective submissions about the purposeful choice situation. There is a range of variables to describe subjective understanding the properties of the choice situation. These variables help to receive linguistic estimates and fuzzy numerical estimates. It allows increasing the degree of adequacy for modeling people’s behavior in different situations and predicting their behavior for different management methods more precisely. Referenses 1. Novikov D.A. Modern problems of the theory of organizational systems managing. Chelovecheskiy faktor v upravlenii [Human factor in management]. Ed. Abramova N.A., Ginsberg K.S., Novikov D.A. Moscow, KomKniga Publ., 2006, pp. 391–407 (in Russ.). 2. Ackoff R., Emery F. On purposeful systems: An interdisciplinary analysis of individual and social behavior as a system of purposeful events. Aldine-Atherton, Chicago, 1972 (Russ. ed.: Rubalskiy G.B., Ushakov I.A. Moscow, Sov. Radio Publ., 1974, 272 p.). 3. Simon Н.А. Information processing theory of human problem solving. (Ed.: Estes W.K.). Handbook of learning and cognitive processes. Hillsdale, N.J., Erlbaurn Publ., 1978, vol. 5. 4. Tversky А., Kahneman D. The framing of decisions and the rationality of choice. Science, New Series. 1981, vol. 211, no. 4481, pp. 453–458 (33, 34). 5. Yarushkina N.G. Osnovy teorii nechyotkikh i gibridnykh sistem [The basics of fuzzy and hybrid systems theory]. Moscow, Finansy i statistika Publ., 2004 (in Russ.). 6. Burkov V.N., Vinogradov G.P., Kuznetsov V.N., Pa- lyukh B.V., Semenov N.A. Intelligent active systems. Trudy 12 nats. konf. po iskusstvennomu intellektu s mezhd. uchastiem [Proc. of the 12 national conf. on Artificial Intelligence with int. engagement]. Moscow, Fizmatlit Publ., 2010, vol. 3, pp. 35–43 (in Russ.). 7. Vinogradov G.P. Simulation of the representations forming by an intelligent agent in the purposeful choice state. Problemy informatiki [Problems of Computer Science]. 2010, no. 3, pp. 66–72 (in Russ.). 8. Vinogradov G.P., Kuznetsov V.N., Semenov N.A. A mo- del of the agent’s representations and beliefs about the purposeful choice state. Trudy Kongressa po intellektualnym sistemam i informatsionnym tekhnologiyam "AIS-IT'10" [Proc. of Congress on Intelligent Systems and Information Technology "AIS-IT'10"]. Moscow, Fizmatlit Publ., 2010, vol. 1, pp. 119–123 (in Russ.). 9. Novikov D.A. Matematicheskie modeli formirovaniya i funktsionirovaniya komand [Mathematical models of commands’ formation and functioning]. Moscow, Fizmatlit Publ., 2008 (in Russ.). 10. Vinogradov G.P., Kuznetsov V.N. Simulation of the agent’s behavior considering subjective views in the purposeful choice situation. Iskusstvenny intellekt i prinyatie resheniy [Artificial intelligence and decision-making]. 2011, no. 3, pp. 58–72 (in Russ.). 11. Avtonomov V.S. Model cheloveka v ekonomicheskoy nauke [Human model in economical science]. St. Petersburg, Ekonomicheskaya shkola Publ., 1998, 72 p. (in Russ.) 12. Vagin V.N., Pospelov D.A., eds. Dostoverny i pravdopodobny vyvod v intellektualnykh sistemakh [Reliable and realistic output in intelligent systems]. 2nd ed., Moscow, Fizmatlit Publ., 2008, 712 p. (in Russ.). |
http://swsys.ru/index.php?id=3806&lang=.&page=article |
|