4.1 Development of personalized mobile information service of university library based on cloud computing platform
Cloud computing platform has laid a very important technical foundation for the realization of humanized mobile information services in libraries. Cloud computing node is the main core component in the high-performance computer system architecture, except for the acceleration node, master control unit, network, etc. The efficiency of computer nodes has become the key issue to achieve the overall goal of "high efficiency" of the whole machine.
First of all, some large high-performance computer systems in research and use are very concerned about the breakthrough of computing node efficiency. Secondly, in terms of data processor nodes, due to the emergence of multi-core processors in recent years, the main efficient computing nodes have generally used multi-channel multi-core technology. In the current technology, the main routes are 4 and 8, while the higher density routes are relatively lacking. Furthermore, in terms of the selection of data processors, the current node computers of high efficiency cloud computing technology mainly choose foreign data processors.
Table 1
Comparison of cloud computing node machines
Cloud computing node machine | Number and type of CPUs | CPU clock frequency | Node power consumption | Node volume | Multi core CPU |
Cloud computing node with multi-core CPU | 16 way, 4-core Loongson 3A | 1.0GHz | About 300W | 1 U Rack | Y |
IBM Blnde JS20 Server | 2 way, 1BM PowerPC970 | 2.2GHz | About 395W | About 1/2 U Blade | N |
IBM BladeCenter JS22 Experiences server | 4-way, POWER6 | 4.0GHz | About 350W | About 1/2 U Blade | N |
TYAN GT62B8230-L server | 2-way AMD 12-Core Opteron6100 | 2.1GHz | About 350W | 1 U Rack | N |
In general, the more new historical browsing tags are used, the more attention they will have to the document. Therefore, when using historical browsing tags to judge the "attention" of browsing documents, different link weights should be given to each historical browsing data according to its time stage.
The number of time cycles in the past is represented by T, and FILE represents the number of requested views of file f in the T time cycle. Then the "heat" of document f can be calculated:
$$\text{P}\text{o}\text{p}\text{u}\text{l}\text{a}\text{r}\text{i}\text{t}\text{y}\left(\text{f}\right)=\sum _{\text{t}=1}^{\text{T}}({2}^{(\text{t}-\text{T})}\times {\text{A}}_{\text{f}}^{\text{t}})$$
14
Then the popular file p that needs to create a copy is:
$$\text{p}=\text{m}\text{a}\text{x}\left[\text{P}\text{o}\text{p}\text{u}\text{l}\text{a}\text{r}\text{i}\text{t}\text{y}\right(\text{f}\left)\right]$$
15
Next, you can use the average browsing "heat" of the relative popular file p and the average "heat" of all documents in the collection FILE to determine the number of copies made by all popular files p.
The average heat of all files in the collection FILE is:
$$\stackrel{-}{\text{P}\text{o}\text{p}\text{u}\text{l}\text{a}\text{r}\text{i}\text{t}\text{y}\left(\text{f}\right)}=\frac{\sum _{\text{f}\in \text{F}\text{I}\text{L}\text{E}}\text{P}\text{o}\text{p}\text{u}\text{l}\text{a}\text{r}\text{i}\text{t}\text{y}\left(\text{f}\right)}{\left|\text{F}\text{I}\text{L}\text{E}\right|}$$
16
Where, | FILE | represents the number of files in the collection FILE, the number of copies to be created for popular FILE p is:
$${\text{N}}^{\text{p}}=\left[\frac{\text{P}\text{o}\text{p}\text{u}\text{l}\text{a}\text{r}\text{i}\text{t}\text{y}\left(\text{p}\right)}{\stackrel{-}{\text{P}\text{o}\text{p}\text{u}\text{l}\text{a}\text{r}\text{i}\text{t}\text{y}\left(\text{f}\right)}}\right]$$
17
The number of times P is requested and used in different regions is also different, that is, the contribution to its "popularity" is also different in different regions. By collecting relevant data sent from different nodes, different summary points control the historical use of all data files. Therefore, according to the historical use data of p controlled by the summary point, the "heat" of p use in different regions can be calculated separately:
$${\text{P}\text{o}\text{p}\text{u}\text{l}\text{a}\text{r}\text{i}\text{t}\text{y}}_{\text{r}}\left(\text{p}\right)=\sum _{\text{t}=1}^{\text{T}}\left({2}^{\left(\text{t}-\text{T}\right)}\times {\text{A}}_{\text{p}}^{\text{t}}\right),\text{r}=\text{1,2},\dots ,\text{R}$$
18
In the case of meeting the constraints such as the QoS standard of the task scheduling system, the cloud service provider's fees are mainly computing costs. Assuming that the application domain target ti will be allocated to the computing resource ci and the matrix element xi is configured, the computing cost to be paid by the node cj to achieve the information scheduling function of the operation target ti is:
$$\text{C}\text{o}\text{m}\text{p}\_{\text{c}\text{o}\text{s}\text{t}}_{\text{j}}^{\text{i}}=\text{I}\text{n}\text{s}\text{t}\text{r}\text{u}\_{\text{c}\text{o}\text{s}\text{t}}_{\text{j}}\times \text{I}\text{n}\text{s}\text{t}\text{r}\text{u}\_{\text{c}\text{o}\text{u}\text{n}\text{t}}_{\text{i}}$$
19
The overhead cost and the number of unit instructions contained in the application task t. Compared with the above calculation overhead cost, the service revenue that the node cj can obtain by completing the scheduling request of the application task ti is:
$${\text{R}\text{e}\text{v}\text{e}\text{n}\text{u}\text{e}}_{\text{j}}^{\text{i}}={\text{B}\text{u}\text{d}\text{g}\text{e}\text{t}}_{\text{i}}-{\text{F}\text{i}\text{n}\text{e}}_{\text{j}}^{\text{i}}$$
20
Further, the total service revenue that the entire cloud computing system can obtain by completing the scheduling operation of all tasks in the application task set to be scheduled is:
$$\text{R}\text{e}\text{v}\text{e}\text{n}\text{u}\text{e}=\sum _{\text{j}=1}^{\text{n}}\sum _{\text{i}=1}^{\text{m}}({\text{R}\text{e}\text{v}\text{e}\text{n}\text{u}\text{e}}_{\text{j}}^{\text{i}}\times {\text{x}}_{\text{i},\text{j}})$$
21
Cloud computing system is a commercial public service. On the premise of meeting the constraint requirements of task scheduling goals, it will spend as little as possible to obtain as much business profits as possible; Compared with fixed computing costs, the more total service profits cloud services expect to obtain, the better.
In recent years, some fine-grained resource scheduling strategies have also emerged on the MapReduce platform. Among them, the Through put Scheduler is designed for heterogeneous MapReduce clusters. It uses machine learning methods to predict the resource requirements of tasks and the computing capacity of MapReduce cluster nodes, and then formulates the optimal task allocation strategy. Polo adjusts the number of task slot slots of MapReduce cluster nodes according to job resource information, and formulates task allocation strategies to maximize the utilization of cluster resources. However, these two scheduling algorithms are still aimed at slot one-dimensional resource scheduling.
The delay scheduling strategy is also applied to the cloud platform of elastic computing, which also uses a distributed file system, such as HDFS, but also faces the problem of data locality. This method uses the delay configuration information for local targets without data to prevent remote reading of targets.
The Genexalized Fairness on Jobs policy fairness measure is based on the number of jobs allocated to each user.
Table 2
Comparison of Typical Scheduling Algorithms
| FIFO | Capacity | Fair | DRF | Asset | CEEI |
Design idea | Submission time sorting | Resource occupancy | Single resource equity | Equity of main resources | Equity of total resources | Maximize resource consumption |
Multidimensional resources | no | no | no | yes | yes | yes |
Big homework | Incremental mechanism | Incremental mechanism | no | no | no | no |
The existing typical scheduling algorithms are shown in Table 2.
4.2 Development of personalized mobile information service of university library based on multi-core processor
Interact with the school educational administration system, and submit relevant information about the current learning content to the user at the appropriate time through the school curriculum. The functions of the personalized mobile information service system of university library mainly include: (1) converting the user's information behavior statistics into the user's information content needs; (2) Collect user information behavior statistics; (3) Generate recommendation information and submit information content service through calculation. With the help of multi-core processors and cloud computing platforms, resources can be integrated. As shown in Table 3, in multi-core processors, each data processor has a buffer that directly corresponds to each other, but they have two conflicting parameters: efficiency and volume. If we only consider the overall hit time of the processor, we must reduce the amount of cache space to reduce access costs; However, if we fully consider increasing the amount of cache space, we must weaken the hit time, which also leads to the emergence of multi-level cache design.
Table 3
AMD X4 processor cache hit time
Cache Level | size | Degree of association | Hit Time |
L1 | 64KB | 2-way | 3 Clock cycles |
L2 | 512KB | 16-way | ≥ 9 clock cycles |
L3 | 2MB | 32-way | About 38 clock cycles |
Although the performance of each level of cache is similar, its structure and performance are different. Generally speaking, all the information in the high-level cache structure includes the low-level cache structure, while there are many differences in the multi-level cache structure. Because the multi-level cache structure should consider the characteristics of multi-level exclusion, and all the information in the high and low caches should not intersect with each other. Therefore, if an error is requested in the high-level cache but hit in the low-level cache, the information in the high and low caches can be exchanged with each other. The advantage of implementing multi-level exclusion is that you can give the low-level cache as much information as possible to save space.
For the detection of quick sort algorithm, unordered data blocks with different information scales are generally used. From the running results, it can be found that the most time-consuming functions generally include the information input module, quick sort algorithm recursion and interval analysis module. The information input module generally includes the reading process of a large number of data files, which has nothing to do with the quick sort algorithm itself, Therefore, the function analysis focuses on the quick sort module and the region analysis module.
As shown in Fig. 3, by analyzing the data of the same model on different data scales, it can be seen that the execution time of the quick sort algorithm accounts for 5% of the total execution time of the calculation, while the execution time of the interval partition model accounts for 90% of the total execution time of the calculation. Through this comparison, it is not difficult to see that the call relationship between the modules in the quick sort algorithm is relatively stable, The bottleneck problem of the algorithm is generally focused on the interval division module and recursive quick sorting module.
Table 4
Multi core synchronization effect test
| Nuclear 0 | Nuclear 1 | Nuclear 2 | Nuclear 3 | Nuclear 4 | Nuclear 5 | Nuclear 6 | Nuclear 7 |
Number of instruction cycles after 50 synchronizations of the first type | 640355 | 640640 | 640679 | 640363 | 640787 | 640853 | 640398 | 640166 |
Number of instruction cycles after 50 synchronizations of the second type | 653340 | 638002 | 653439 | 653432 | 653440 | 653209 | 637654 | 653120 |
Number of instruction cycles after 50 synchronizations of the third type | 653319 | 653239 | 639000 | 653702 | 653290 | 653118 | 653820 | 653833 |
According to Table 4, the multi-core communication network with AMP structure can meet the requirements of millisecond data transmission for transient interference protection. When multiple cores are the same, there are three options: one is to get the semaphore before the algorithm module, and the algorithm module then releases the semaphore to achieve synchronization; Second, it does not need semaphores at all, but synchronizes by setting flag bits in shared memory; Third, when the algorithm module is completed, you can use only the semaphore obtained or only the semaphore released to judge, thus realizing step.
With the support of multi-core processors, the personalized mobile information service platform of university library on cloud computing platform transmits text faster. The multi-core processor can also use the service number to provide the template mode of fast push information notification. On the premise of user approval, the multi-core processor can actively send the user information such as the activity information that the user is borrowing books, the information that the books are about to expire, and the information about the emergency closure of the library. The active push message service mode will provide greater convenience for users.