| Literature DB >> 35585154 |
Pengzhan Guo1, Keli Xiao2, Zeyang Ye3, Hengshu Zhu4, Wei Zhu5.
Abstract
Career planning consists of a series of decisions that will significantly impact one's life. However, current recommendation systems have serious limitations, including the lack of effective artificial intelligence algorithms for long-term career planning, and the lack of efficient reinforcement learning (RL) methods for dynamic systems. To improve the long-term recommendation, this work proposes an intelligent sequential career planning system featuring a career path rating mechanism and a new RL method coined as the stochastic subsampling reinforcement learning (SSRL) framework. After proving the effectiveness of this new recommendation system theoretically, we evaluate it computationally by gauging it against several benchmarks under different scenarios representing different user preferences in career planning. Numerical results have demonstrated that our system is superior to other benchmarks in locating promising optimal career paths for users in long-term planning. Case studies have further revealed that our SSRL career path recommendation system would encourage people to gradually improve their career paths to maximize long-term benefits. Moreover, we have shown that the initial state (i.e., the first job) can have a significant impact, positively or negatively, on one's career, while in the long-term view, a carefully planned career path following our recommendation system may mitigate the negative impact of a lackluster beginning in one's career life.Entities:
Mesh:
Year: 2022 PMID: 35585154 PMCID: PMC9117248 DOI: 10.1038/s41598-022-11872-8
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.996
Figure 1The SSRL framework. The inputs of SSRL contain user-provided information, including the current employer and position type as well as the optional working duration and work history information. Then, a four-step iteration handles the optimization process to provide personalized career guidance. Step 1 initializes the process and stochastically generates employer subsample based on the corresponding user states during the iteration. Step 2 handles RL environment construction and further performs the RL to explore optimal policy and generate the best career path based on the subsample. Note that the path evaluation function guiding the policy exploration jointly considers company and position features along with user preference and potential work experience gain in their career life. Step 3 determines whether to accept the current best path. To avoid being trapped in local optima, a cool-down strategy is proposed to allow accepting worse cases according to a probability following the Boltzmann distribution. Step 4 updates the candidate state accordingly and loops back to Step 1 for new subsampling. Once the terminating condition is met, SSRL will output the recommended career path.
Figure 2Average path score (per month) and accumulative reward. In this figure, (a)–(d) plot the average path scores (per month); (e)–(h) plot the accumulative rewards of recommended career paths over a 20-year career timeline (). Four experimental scenarios are tested. In Scenario 1, we consider a general case where people have no specific preference (i.e., the weight set of company-related features [reputation, popularity, average staying duration, smooth transfer rate] is set to [0.25, 0.25, 0.25, 0.25]. Scenario 2 represents a case where users may have a specific preference on one of the features, e.g., the reputation. The weight is set to [0.7, 0.1, 0.1, 0.1]. Scenario 3 mimics a dynamic preference in someone’s career life. We assume someone has a specific preference on the company reputation in the first 10 years (the same as Scenario 2), then the weight is updated to [0.1, 0.1, 0.7, 0.1], representing a preference change. Scenario 4 demonstrates an extreme case that someone has a preference on specific companies at a specific time. For example, someone has a clear plan to join Bank of Boston (or similar companies) at the early stage in her career (e.g., we lock our recommendations to the preferred companies at the third decision point [around the sixth career year]). For each case, we simulate the path with 30 random initial states, and the mean values are then plotted along with their standard errors [the error bars in (a)–(d)].
Figure 3Career guidance. This figure illustrates career paths recommended by different methods, given the initial situation of an individual in the Navel Group as an engineer and with no specific user preference. We simulate 20-year career plans based on our method (SSRL) and five baseline methods. Compared to all baselines, SSRL shows a significant advantage in attaining the best-quality career path according to the pre-defined quality score. Also, SSRL recommended a gradually improved path, while all other baselines resulted in fluctuated quality of job mobility. If promotion and re-education information is not considered, SSRL tends to recommend the same-type of positions, while the companies may be from different industries.
Top-10 most frequent recommendations via SSRL.
| Preferred company/organization feature | ||||
|---|---|---|---|---|
| Rank | Reputation | Popularity | Average staying duration | Smooth transfer rate |
| 1 | Microsoft | IBM | HM Forces | Bank of America |
| 2 | IBM | Canon | Royal Air Force | IBM |
| 3 | Monsanto | Hewlett Packard | Indian Air Force | Pfizer |
| 4 | Pfizer | Accenture | U.S. Navy | Nortel |
| 5 | Walgreens Boots Alliance | Boys and Girls Clubs of America | Royal Navy | JPMorgan Chase |
| 6 | JPMorgan Chase | Bank of America | Royal Australian Navy | Northwest Airlines |
| 7 | Alcoa | Ernst and Young | Northwest Airlines | Lockheed Martin |
| 8 | General Motors | Microsoft | Royal Australian Air Force | Level 3 Comm. |
| 9 | UnitedHealth | JPMorgan Chase | U.S. Air Force | Sun |
| 10 | BMS | PwC | British Army | UnitedHealth |
This table reports the most frequently recommended companies or organizations via our SSRL framework, based on different user preferences on the four company-specific features in our evaluation model. *Given that the “average staying duration” is the preferred feature, the top-10 recommendations obtained include many military positions. Military is not a sector defined by GICS. Thus, we consider it as a spacial class. If removing military positions, the top-10 recommendations are: Northwest Airlines, Electronic Data Systems, Various Inc., Delta Air Lines, Kodak, Texas Instruments, Carlson Marketing, Bell Laboratories, Ford Motor Credit Company, CGG.
Figure 4Business sectors of the top recommendations. This figure illustrates the portion of business sector of the top recommendations under different user preferences. The results are based on 200 career path simulations under different initial states. The weight for the preferred feature is set to be 0.7 while the rest being 0.1 each. We follow the Global Industry Classification Standard (GICS) to determine the business sectors.
Career path planning based on different initial companies and positions.
| Initial position type | ||||
|---|---|---|---|---|
| Initial company [rating] | Legal | Sales | Engineering | Support |
| Panel A: Average path score | ||||
| Barceló [34.00] | 60.34 (0.65) | 59.48 (0.50) | 60.30 (0.52) | 59.33 (0.59) |
| ACC | 59.92 (0.51) | 60.09 (0.52) | 60.60 (0.54) | 59.31 (0.50) |
| AstraZeneca [62.35] | 67.52 (0.52) | 66.51 (0.51) | 68.62 (0.45) | 67.54 (0.50) |
| FedEx Office [62.44] | 68.77 (0.33) | 67.28 (0.63) | 68.05 (0.46) | 67.29 (0.58) |
| Panel B: Good path percentage | ||||
| Barceló [34.00] | 0.03 | 0.00 | 0.00 | 0.03 |
| ACC | 0.00 | 0.00 | 0.00 | 0.00 |
| AstraZeneca [62.35] | 0.57 | 0.60 | 0.73 | 0.67 |
| FedEx Office [62.44] | 0.87 | 0.60 | 0.70 | 0.60 |
This table summarizes the average path scores and good path percentage based on experiments with different initial companies and position types. We selected four companies (two average-rating companies and two high-rating companies) and four position types, leading to 16 combinations for the investigation. For each combination, 30 independent experiments were conducted. Panel A reports the average path score of the recommended career paths, along with corresponding standard errors (in parentheses). Panel B reports the “good path” percentages based on the same experiments. We define a good career path as one with a score greater than 66.62. *ACC: American Campus Communities Inc.
Comparison of benchmark methods.
| Method | Description | Advantage | Disadvantage |
|---|---|---|---|
| JBMUL | Traditional greedy method | Efficient for convex situation | Trap into local optimal easily in non-convex situation |
| IGM | Modified greedy method | Available to leave local optimal | Convergence rate is slow |
| MGM | Modified greedy method | Available to leave local optimal | Performance is unstable |
| TTD | Traditional RL method based on Q-table | Efficient for long-term sequence decision | Limited by the size of actions and states |
| PDQN | Advanced RL framework based on neural network | Efficient for long-term sequence decision with infinite time. | Limited by the size of actions |