首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 46 毫秒
1.
Software Distributed Shared Memory (DSM) systems can be used to provide a coherent shared address space on multicomputers and other parallel systems without support for shared memory in hardware. The coherency software automatically translates shared memory accesses to explicit messages exchanged among the nodes in the system. Many applications exhibit a good performance on such systems but it has been shown that, for some applications, performance critical messages can be delayed behind less important messages because of the enqueuing behavior in the communication libraries used in current systems. We present in this paper a new portable communication library that supports priorities to remedy this situation. We describe an implementation of the communication library and a quantitative model that is used to estimate the performance impact of priorities for a typical situation. Using the model, we show that the use of high-priority communication reduces the latency of performance critical messages substantially over a wide range of network design parameters. The latency is reduced with up to 10–25% for each delaying low priority message in the queue ahead.  相似文献   

2.
In the large-scale parallel computing environment, resource allocation and energy efficient techniques are required to deliver the quality of services (QoS) and to reduce the operational cost of the system. Because the cost of the energy consumption in the environment is a dominant part of the owner’s and user’s budget. However, when considering energy efficiency, resource allocation strategies become more difficult, and QoS (i.e., queue time and response time) may violate. This paper therefore is a comparative study on job scheduling in large-scale parallel systems to: (a) minimize the queue time, response time, and energy consumption and (b) maximize the overall system utilization. We compare thirteen job scheduling policies to analyze their behavior. A set of job scheduling policies includes (a) priority-based, (b) first fit, (c) backfilling, and (d) window-based policies. All of the policies are extensively simulated and compared. For the simulation, a real data center workload comprised of 22385 jobs is used. Based on results of their performance, we incorporate energy efficiency in three policies i.e., (1) best result producer, (2) average result producer, and (3) worst result producer. We analyze the (a) queue time, (b) response time, (c) slowdown ratio, and (d) energy consumption to evaluate the policies. Moreover, we present a comprehensive workload characterization for optimizing system’s performance and for scheduler design. Major workload characteristics including (a) Narrow, (b) Wide, (c) Short, and (d) Long jobs are characterized for detailed analysis of the schedulers’ performance. This study highlights the strengths and weakness of various job scheduling polices and helps to choose an appropriate job scheduling policy in a given scenario.  相似文献   

3.
In this paper, researching on task scheduling is a way from the perspective of resource allocation and management to improve performance of Hadoop system. In order to save the network bandwidth resources in Hadoop cluster environment and improve the performance of Hadoop system, a ReduceTask scheduling strategy that based on data-locality is improved. In MapReduce stage, there are two main data streams in cluster network, they are slow task migration and remote copies of data. The two overlapping burst data transfer can easily become bottlenecks of the cluster network. To reduce the amount of remote copies of data, combining with data-locality, we establish a minimum network resource consumption model (MNRC). MNRC is used to calculate the network resources consumption of ReduceTask. Based on this model, we design a delay priority scheduling policy for the ReduceTask which is based on the cost of network resource consumption. Finally, MNRC is verified by simulation experiments. Evaluation results show that MNRC outperforms the saving cluster network resource by an average of 7.5% in heterogeneous.  相似文献   

4.
Many-task computing aims to bridge the gap between two computing paradigms, high throughput computing and high performance computing. Many-task computing denotes high-performance computations comprising multiple distinct activities, coupled via file system operations. The aggregate number of tasks, quantity of computing, and volumes of data may be extremely large. Traditional techniques found in production systems in the scientific community to support many-task computing do not scale to today’s largest systems, due to issues in local resource manager scalability and granularity, efficient utilization of the raw hardware, long wait queue times, and shared/parallel file system contention and scalability. To address these limitations, we adopted a “top-down” approach to building a middleware called Falkon, to support the most demanding many-task computing applications at the largest scales. Falkon (Fast and Light-weight tasK executiON framework) integrates (1) multi-level scheduling to enable dynamic resource provisioning and minimize wait queue times, (2) a streamlined task dispatcher able to achieve orders-of-magnitude higher task dispatch rates than conventional schedulers, and (3) data diffusion which performs data caching and uses a data-aware scheduler to co-locate computational and storage resources. Micro-benchmarks have shown Falkon to achieve over 15K+ tasks/s throughputs, scale to hundreds of thousands of processors and to millions of queued tasks, and execute billions of tasks per day. Data diffusion has also shown to improve applications scalability and performance, with its ability to achieve hundreds of Gb/s I/O rates on modest sized clusters, with Tb/s I/O rates on the horizon. Falkon has shown orders of magnitude improvements in performance and scalability than traditional approaches to resource management across many diverse workloads and applications at scales of billions of tasks on hundreds of thousands of processors across clusters, specialized systems, Grids, and supercomputers. Falkon’s performance and scalability have enabled a new class of applications called Many-Task Computing to operate at previously so-believed impossible scales with high efficiency.  相似文献   

5.
The current works about MapReduce task scheduling with deadline constraints neither take the differences of Map and Reduce task, nor the cluster’s heterogeneity into account. This paper proposes an extensional MapReduce Task Scheduling algorithm for Deadline constraints in Hadoop platform: MTSD. It allows user specify a job’s deadline and tries to make the job be finished before the deadline. Through measuring the node’s computing capacity, a node classification algorithm is proposed in MTSD. This algorithm classifies the nodes into several levels in heterogeneous clusters. Under this algorithm, we firstly illuminate a novel data distribution model which distributes data according to the node’s capacity level respectively. The experiments show that the node classification algorithm can improved data locality observably to compare with default scheduler and it also can improve other scheduler’s locality. Secondly, we calculate the task’s average completion time which is based on the node level. It improves the precision of task’s remaining time evaluation. Finally, MTSD provides a mechanism to decide which job’s task should be scheduled by calculating the Map and Reduce task slot requirements.  相似文献   

6.

Non-orthogonal multiple access (NOMA) along with cognitive radio (CR) have been recently configured as potential solutions to fulfill the extraordinary demands of the fifth generation (5G) and beyond (B5G) networks and support the Internet of Thing (IoT) applications. Multiple users can be served within the same orthogonal domains in NOMA via power-domain multiplexing, whilst CR allows secondary users (SUs) to access the licensed spectrum frequency. This work investigates the possibility of combining orthogonal frequency division multiple access (OFDMA), NOMA, and CR, referred to as hybrid OFDMA-NOMA CR network. With this hybrid technology, the licensed frequency is divided into several channels, such as a group SUs is served in each channel based on NOMA technology. In particular, a rate-maximization framework is developed, at which user pairing at each channel, power allocations for each user, and secondary users activities are jointly considered to maximize the sum-rate of the hybrid OFDMA-NOMA CR network, while maintaining a set of relevant NOMA and CR constraints. The developed sum-rate maximization framework is NP-hard problem, and cannot be solved through classical approaches. Accordingly, we propose a two-stage approach; in the first stage, we propose a novel user pairing algorithm. With this, an iterative algorithm based on the sequential convex approximation is proposed to evaluate the solution of the non-convex rate-maximization problem, in the second stage. Results show that our proposed algorithm outperforms the existing schemes, and CR network features play a major role in deciding the overall network’s performance.

  相似文献   

7.
MapReduce offers an ease-of-use programming paradigm for processing large data sets, making it an attractive model for opportunistic compute resources. However, unlike dedicated resources, where MapReduce has mostly been deployed, opportunistic resources have significantly higher rates of node volatility. As a consequence, the data and task replication scheme adopted by existing MapReduce implementations is woefully inadequate on such volatile resources. In this paper, we propose MOON, short for MapReduce On Opportunistic eNvironments, which is designed to offer reliable MapReduce service for opportunistic computing. MOON adopts a hybrid resource architecture by supplementing opportunistic compute resources with a small set of dedicated resources, and it extends Hadoop, an open-source implementation of MapReduce, with adaptive task and data scheduling algorithms to take advantage of the hybrid resource architecture. Our results on an emulated opportunistic computing system running atop a 60-node cluster demonstrate that MOON can deliver significant performance improvements to Hadoop on volatile compute resources and even finish jobs that are not able to complete in Hadoop.  相似文献   

8.
Shao  Bilin  Ji  Yanyan 《Cluster computing》2021,24(3):1989-2000

In recent years, how to design efficient auditing protocol to verify the integrity of users’ data, which is stored in cloud services provider (CSP), becomes a research focus. Homomorphic message authentication code (MAC) and homomorphic signature are two popular techniques to respectively design private and public auditing protocols. On the one hand, it is not suitable for the homomorphic-MAC-based auditing protocols to be outsourced to third-party auditor (TPA), who has more professional knowledge and computational abilities, although they have high efficiencies. On the other hand, the homomorphic-signature-based ones are very suitable for employing TPA without compromising user’s signing key but have very low efficiency (compared to the former case). In this paper, we propose a new auditing protocol, which perfectly combines the advantages of above two cases. In particular, it is almost as efficient as a homomorphic-MAC-based protocol proposed by Zhang et al. recently. Moreover, it is also suitable for outsourcing to TPA because it does not compromise the privacy of users’ signing key, which can be seen from our security analysis. Finally, numerical analysis and experimental results demonstrate the high-efficiency of our protocol.

  相似文献   

9.
Developing an efficient parallel application is not an easy task, and achieving a good performance requires a thorough understanding of the program’s behavior. Careful performance analysis and optimization are crucial. To help developers or users of these applications to analyze the program’s behavior, it is necessary to provide them with an abstraction of the application performance. In this paper, we propose a dynamic performance abstraction technique, which enables the automated discovery of causal execution paths, composed of communication and computational activities, in MPI parallel programs. This approach enables autonomous and low-overhead execution monitoring that generates performance knowledge about application behavior for the purpose of online performance diagnosis. Our performance abstraction technique reflects an application behavior and is made up of elements correlated with high-level program structures, such as loops and communication operations. Moreover, it characterizes all elements with statistical execution profiles. We have evaluated our approach on a variety of scientific parallel applications. In all scenarios, our online performance abstraction technique proved effective for low-overhead capturing of the program’s behavior and facilitated performance understanding.  相似文献   

10.
The delivery of scalable, rich multimedia applications and services on the Internet requires sophisticated technologies for transcoding, distributing, and streaming content. Cloud computing provides an infrastructure for such technologies, but specific challenges still remain in the areas of task management, load balancing, and fault tolerance. To address these issues, we propose a cloud-based distributed multimedia streaming service (CloudDMSS), which is designed to run on all major cloud computing services. CloudDMSS is highly adapted to the structure and policies of Hadoop, thus it has additional capacities for transcoding, task distribution, load balancing, and content replication and distribution. To satisfy the design requirements of our service architecture, we propose four important algorithms: content replication, system recovery for Hadoop distributed multimedia streaming, management for cloud multimedia management, and streaming resource-based connection (SRC) for streaming job distribution. To evaluate the proposed system, we conducted several different performance tests on a local testbed: transcoding, streaming job distribution using SRC, streaming service deployment and robustness to data node and task failures. In addition, we performed three different tests in an actual cloud computing environment, Cloudit 2.0: transcoding, streaming job distribution using SRC, and streaming service deployment.  相似文献   

11.
Cheng  Xiaoming  Wang  Lei  Zhang  Pengchao  Wang  Xinkuan  Yan  Qunmin 《Cluster computing》2022,25(3):2107-2123

Household electricity consumption has been rising gradually with the improvement of living standards. Making short-term load forecasting at the small-scale users plays an increasingly important role in the future power network planning and operation. To meet the efficiency of the dispatching system and the demand of human daily power consumption, an optimal forecasting model Attention-CNN-GRU of small-scale users load at various periods of the day based on family behavior pattern recognition is proposed in this study. The low-level data information (smart meter data) is used to build the high-level model (small-scale users load). Attention mechanism and convolutional neural networks (CNN) can further enhance the prediction accuracy of gated recurrent unit (GRU) and notably shorten its prediction time. The recognition of family behavior patterns can be achieved through the users’ smart meter data, and users are aggregated into K categories. The results of optimal K category prediction under the family behavior model are summarized as the final prediction outcome. This idea framework is tested on real users’ smart meter data, and its performance is comprehensively compared with different benchmarks. The results present strong compatibility in the small-scale users load forecasting model at various periods of the day and swift short-term prediction of users load compared to other prediction models. The time is shortened by 1/4 compared with the GRU/LSTM model. Furthermore, the accuracy is improved to 92.06% (MAPE is 7.94%).

  相似文献   

12.
In this paper, we propose a new framework to analyze the temporal dynamics of the emotional stimuli. For this framework, both electroencephalography signal and visual information are of great importance. The fusion of visual information with brain signals allows us to capture the users’ emotional state. Thus we adopt previously proposed fuzzy-GIST as emotional feature to summarize the emotional feedback. In order to model the dynamics of the emotional stimuli sequence, we develop a recurrent neuro-fuzzy network for modeling the dynamic events of emotional dimensions including valence and arousal. It can incorporate human expertise by IF-THEN fuzzy rule while recurrent connections allow the fuzzy rules of network to see its own previous output. The results show that such a framework can interact with human subjects and generate arbitrary emotional sequences after learning the dynamics of an emotional sequence with enough number of samples.  相似文献   

13.
Cumulative effect in social contagion underlies many studies on the spread of innovation, behavior, and influence. However, few large-scale empirical studies are conducted to validate the existence of cumulative effect in information diffusion on social networks. In this paper, using the population-scale dataset from the largest Chinese microblogging website, we conduct a comprehensive study on the cumulative effect in information diffusion. We base our study on the diffusion network of message, where nodes are the involved users and links characterize forwarding relationship among them. We find that multiple exposures to the same message indeed increase the possibility of forwarding it. However, additional exposures cannot further improve the chance of forwarding when the number of exposures crosses its peak at two. This finding questions the cumulative effect hypothesis in information diffusion. Furthermore, to clarify the forwarding preference among users, we investigate both structural motif in the diffusion network and temporal pattern in information diffusion process. Findings provide some insights for understanding the variation of message popularity and explain the characteristics of diffusion network.  相似文献   

14.
Flexible Assembly Systems (FASs), which form an important subset of modern manufacturing systems, are finding increasing use in today's industry. In the planning and design phase of these systems, it is useful to have tools that predict system performance for various operating conditions. In this article, we present such a performance analysis tool based on queueing approximation for a class of FASs, namely, closed-loop flexible assembly systems (CL-FASs). For CL-FASs, we describe iterative algorithms for computing steady-state performance measures, including production rate and station utilizations. These algorithms are computationally simple and have a fast convergence rate. We derive a new approximation to correct the mean delay at each queue. This improves the accuracy of performance prediction, especially in the case of small CL-FASs. Comparisons with simulation results indicate that the approximation technique is reasonably accurate for a broad range of parameter values and system sizes. This makes possible efficient (fast and computationally inexpensive) analysis of CL-FASs under various conditions.  相似文献   

15.

In recent years, cloud computing can be considered an emerging technology that can share resources with users. Because cloud computing is on-demand, efficient use of resources such as memory, processors, bandwidth, etc., is a big challenge. Despite the advantages of cloud computing, sometimes it is not a proper choice due to its delay in responding appropriately to existing requests, which led to the need for another technology called fog computing. Fog computing reduces traffic and time lags by expanding cloud services to the network and closer to users. It can schedule resources with higher efficiency and utilize them to impact the user's experience dramatically. This paper aims to survey some studies that have been done in the field of scheduling in fog/cloud computing environments. The focus of this survey is on published studies between 2015 and 2021 in journals or conferences. We selected 71 studies in a systematic literature review (SLR) from four major scientific databases based on their relation to our paper. We classified these studies into five categories based on their traced parameters and their focus area. This classification comprises 1—performance 2—energy efficiency, 3—resource utilization, 4—performance and energy efficiency, and 5—performance and resource utilization simultaneously. 42.3% of the studies focused on performance, 9.9% on energy efficiency, 7.0% on resource utilization, 21.1% on both performance and energy efficiency, and 19.7% on both performance and resource utilization. Finally, we present challenges and open issues in the resource scheduling methods in fog/cloud computing environments.

  相似文献   

16.
Marcus Hunter and Zandria Robinson have provided us with an innovative methodology for analysing “maps” of African Americans’ lived political, social and cultural experiences past and present. Their powerful innovation is creating maps of the black experience based on black people’s lived experience. They argue that carefully tracing spatially blacks’ political, social and cultural patterns over time leads one to the conclusion that the shared experiences of blacks throughout the polity (and indeed throughout the Diaspora) have far more in common than not. I argue that we can better understand not only the maps that black people have created, but also chocolate cities themselves by using an analytical framework that integrates the analysis of the lived experience of black people with a structural analysis that interrogates the articulation of white supremacy, capitalism and patriarchy. This framework is related to the developing research into racial capitalism that is now being conducted globally.  相似文献   

17.
Large quantities of data have been generated from multiple sources at exponential rates in the last few years. These data are generated at high velocity as real time and streaming data in variety of formats. These characteristics give rise to challenges in its modeling, computation, and processing. Hadoop MapReduce (MR) is a well known data-intensive distributed processing framework using the distributed file system (DFS) for Big Data. Current implementations of MR only support execution of a single algorithm in the entire Hadoop cluster. In this paper, we propose MapReducePack (MRPack), a variation of MR that supports execution of a set of related algorithms in a single MR job. We exploit the computational capability of a cluster by increasing the compute-intensiveness of MapReduce while maintaining its data-intensive approach. It uses the available computing resources by dynamically managing the task assignment and intermediate data. Intermediate data from multiple algorithms are managed using multi-key and skew mitigation strategies. The performance study of the proposed system shows that it is time, I/O, and memory efficient compared to the default MapReduce. The proposed approach reduces the execution time by 200% with an approximate 50% decrease in I/O cost. Complexity and qualitative results analysis shows significant performance improvement.  相似文献   

18.
Neuroprosthetic brain-computer interfaces are systems that decode neural activity into useful control signals for effectors, such as a cursor on a computer screen. It has long been recognized that both the user and decoding system can adapt to increase the accuracy of the end effector. Co-adaptation is the process whereby a user learns to control the system in conjunction with the decoder adapting to learn the user''s neural patterns. We provide a mathematical framework for co-adaptation and relate co-adaptation to the joint optimization of the user''s control scheme ("encoding model") and the decoding algorithm''s parameters. When the assumptions of that framework are respected, co-adaptation cannot yield better performance than that obtainable by an optimal initial choice of fixed decoder, coupled with optimal user learning. For a specific case, we provide numerical methods to obtain such an optimized decoder. We demonstrate our approach in a model brain-computer interface system using an online prosthesis simulator, a simple human-in-the-loop pyschophysics setup which provides a non-invasive simulation of the BCI setting. These experiments support two claims: that users can learn encoders matched to fixed, optimal decoders and that, once learned, our approach yields expected performance advantages.  相似文献   

19.
In recent years, China has experienced rapid economic development and thousands of people escaped from poverty. However, this high-speed development has also led to increased pressure on the environment. Although the Chinese government has focused on solving environmental problems for the past few decades, it appears to have had little effect. Environmental performance evaluation (EPE) is a quantifiable environmental management tool that can evaluate environmental protection effects and provide guidance to improve government efficiency. We use EPE to evaluate China’s environmental performance at the provincial level between 2006 and 2011. In this study, the theme framework and driver force-pressure-state-impact-response (DPSIR) framework models are used to build the composite index (CI) for environmental performance. This index includes 39 indicators in four categories that were selected based on data that can be acquired from China’s Statistical Bureau. The results indicate that the environmental performance index (EPI) of 30 provincial administrative regions (PARs) from 2006 to 2011 ranges from 44.12 (Shanxi, 2006) to 80.87 (Beijing, 2010), from poor to good, respectively. To help develop more effective policies to improve China’s regional environmental performance, cluster analysis (CLA) is applied to divide the 30 PARs into 3 sub-regions. Recommendations for improving the environmental performance of different sub-regions are made to help guide the Chinese government to adjust environmental governance approaches to local conditions.  相似文献   

20.
With the development of IT convergence technologies, users can now more easily access useful information. These days, diverse and far-reaching information is being rapidly produced and distributed instantly in digitized format. Studies are continuously seeking to develop more efficient methods of delivering information to a greater number of users. Image filtering, which extracts features of interest from images, was developed to address the weakness of collaborative filtering, which is limited to superficial data analysis. However, image filtering has its own weakness of requiring complicated calculations to obtain the similarity between images. In this study, to resolve these problems, we propose associative image filtering based on the mining method utilizing the harmonic mean. Using data mining’s Apriori algorithm, this study investigated the association among preferred images from an associative image group and obtained a prediction based on user preference mean. In so doing, we observed a positive relationship between the various image preferences and the various distances between images’ color histograms. Preference mean was calculated based on the arithmetic mean, geometric mean, and harmonic mean. We found through performance analysis that the harmonic mean had the highest accuracy. In associative image filtering, we used the harmonic mean in order to anticipate preferences. In testing accuracy with MAE utilizing the proposed method, this study demonstrated an improvement of approximately 12 % on average compared to previous collaborative image filtering.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号