

{"id":427,"date":"2015-01-02T12:18:12","date_gmt":"2015-01-02T11:18:12","guid":{"rendered":"https:\/\/project.inria.fr\/dalhis\/?page_id=427"},"modified":"2017-11-19T16:13:23","modified_gmt":"2017-11-19T15:13:23","slug":"workplan-2015","status":"publish","type":"page","link":"https:\/\/project.inria.fr\/dalhis\/research\/workplan-2015\/","title":{"rendered":"Workplan 2015"},"content":{"rendered":"<p>We have identifi\fed three major work directions to be investigated as part of DALHIS associated team for 2015:<\/p>\n<p><span style=\"color: #ff6600;\"><strong>Scienti\ffic Workows<\/strong><\/span><br \/>\nThe development of the integrated workfow engine using HOCL and Tigres continues. We recently started to explore the expressiveness of HOCL to express self-adaptive behaviors.\u00a0 Tigres will speci\ffically focus next year on failure recovery and fault tolerance API. Additionally, Tigres will investigate decentralized execution<br \/>\nand optimizations to enable executions on next-generation HPC systems that have deeper I\/O-memory hierarchies. Our work plan is organized as follows.<br \/>\n<strong>Myriads<\/strong><\/p>\n<ul>\n<li>\u000f Complete the implementation of the whole set of templates in HOCL-WMS (2 templates already done)<\/li>\n<\/ul>\n<ul>\n<li>\u000f Produce logs compliant with the Tigres format<\/li>\n<\/ul>\n<ul>\n<li>\u000f Study fault tolerance and recovery mechanisms on top of the HOCL-TS\/Tigres integration<\/li>\n<\/ul>\n<ul>\n<li>\u000f Release HOCL-WMS in open source<\/li>\n<\/ul>\n<p><strong> LBNL<\/strong><\/p>\n<ul>\n<li>\u000f Support failure recovery and repeated executions<\/li>\n<\/ul>\n<ul>\n<li>\u000f Support failure tolerance through the API<\/li>\n<\/ul>\n<ul>\n<li>\u000f Evaluate the need to include loops into Tigres library.<\/li>\n<\/ul>\n<p><strong> Joint work<\/strong><\/p>\n<ul>\n<li>\u000f Develop and evaluate workfow examples such as Montage, MODIS, Light source workflows in the integrated system<\/li>\n<\/ul>\n<ul>\n<li>\u000f Validate a larger set of workfows expressed in Tigres<\/li>\n<\/ul>\n<ul>\n<li>\u000f Evaluate the system using large-scale experiments on HPC and cloud testbeds<\/li>\n<\/ul>\n<p><span style=\"color: #ff6600;\"><strong>Energy-effi\u000ecient cloud elasticity for data-driven applications<\/strong><\/span><br \/>\nDistributed and parallel systems o\u000ber to users tremendous computing capacities. They rely on distributed<br \/>\ncomputing resources linked by networks. They require algorithms and protocols to manage these resources<br \/>\nin a transparent way for users. Recently, the maturity of virtualization techniques has allowed for<br \/>\nthe emergence of virtualized infrastructures (Clouds). These infrastructures provide resources to users<br \/>\ndynamically, and adapted to their needs. By benefi\fting from economies of scale, Clouds can effi\u000eciently<br \/>\nmanage and off\u000ber virtually unlimited numbers of resources, reducing the costs for users.<br \/>\nHowever, the rapid growth for Cloud demands leads to a preoccupying and uncontrolled increase of<br \/>\ntheir electric consumption. In this context, we will focus on data driven applications which require to<br \/>\nprocess large amounts of data. These applications have elastic needs in terms of computing resources<br \/>\nas their workload varies over time. While reducing energy consumption and improving performance are<br \/>\northogonal goals, this internship aims at studying possible trade-o\u000bffs for energy-effi\u000ecient data processing<br \/>\nwithout performance impact. As elasticity comes at a cost of recon\fgurations, these trade-off\u000bs will<br \/>\nconsider the time and energy required by the infrastructure to dynamically adapt the resources to the<br \/>\napplication needs.<br \/>\nThe validations of the proposed algorithms may rely on the French experimental platform named<br \/>\nGrid&#8217;5000. This platform comprises about 8,000 cores geographically distributed in 10 sites linked with<br \/>\na dedicated gigabit network. Some of these sites have wattmeters which provide the consumption of the<br \/>\ncomputing nodes in real-time. This validation step is essential as it will ensure that the selected criteria<br \/>\nare well observed: energy-effi\u000eciency, performance and elasticity.<br \/>\n<strong><span style=\"color: #ff6600;\">Data Ecosystem<\/span><\/strong><br \/>\nScienti\fc now routinely generates large and complex datasets as the result of experiments, observations,<br \/>\nor simulations and the number of scientists analyzing these datasets are also growing. These datasets<br \/>\nand their relationships have become increasingly di\u000efficult to manage and analyze. We will continue our<br \/>\nwork in identifying, developing and implementing the data ecosystem for scienti\ffic applications.<br \/>\nIn addition to the above activities, we will continue our work on FRIEDA to explore elasticity of<br \/>\ndata and its impact on execution. Elasticity and auto-scaling of compute resources has been explored<br \/>\nbefore. However, its interaction with storage and data management is not well understood. There are a<br \/>\nnumber of open issues in this space. First, it is unclear how growing or shrinking data volumes should<br \/>\nbe managed in virtualized environments. Also, when a compute resource is brought up or removed from<br \/>\nthe pool, how should the data on these volumes be managed. In the context of FRIEDA, we will design,<br \/>\nimplement and evaluate data management strategies to manage elasticity. This work will be evaluated<br \/>\non Amazon, Grid 5000 and other cloud testbeds.<br \/>\nThe DALHIS associated team has been in deep discussions about the appropriate data ecosystem<br \/>\nand its components suitable for scienti\fc applications. In the coming year, we will be writing the paper<br \/>\noutlining what a data ecosystem accounts for.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>We have identifi\fed three major work directions to be investigated as part of DALHIS associated team for 2015: Scienti\ffic Workows The development of the integrated workfow engine using HOCL and Tigres continues. We recently started to explore the expressiveness of HOCL to express self-adaptive behaviors.\u00a0 Tigres will speci\ffically focus next year on failure recovery and &hellip; <\/p>\n<p><a class=\"more-link btn\" href=\"https:\/\/project.inria.fr\/dalhis\/research\/workplan-2015\/\">Continue reading<\/a><\/p>\n","protected":false},"author":267,"featured_media":0,"parent":267,"menu_order":6,"comment_status":"closed","ping_status":"closed","template":"","meta":{"footnotes":""},"class_list":["post-427","page","type-page","status-publish","hentry","nodate","item-wrap"],"_links":{"self":[{"href":"https:\/\/project.inria.fr\/dalhis\/wp-json\/wp\/v2\/pages\/427","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/project.inria.fr\/dalhis\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/project.inria.fr\/dalhis\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/project.inria.fr\/dalhis\/wp-json\/wp\/v2\/users\/267"}],"replies":[{"embeddable":true,"href":"https:\/\/project.inria.fr\/dalhis\/wp-json\/wp\/v2\/comments?post=427"}],"version-history":[{"count":9,"href":"https:\/\/project.inria.fr\/dalhis\/wp-json\/wp\/v2\/pages\/427\/revisions"}],"predecessor-version":[{"id":531,"href":"https:\/\/project.inria.fr\/dalhis\/wp-json\/wp\/v2\/pages\/427\/revisions\/531"}],"up":[{"embeddable":true,"href":"https:\/\/project.inria.fr\/dalhis\/wp-json\/wp\/v2\/pages\/267"}],"wp:attachment":[{"href":"https:\/\/project.inria.fr\/dalhis\/wp-json\/wp\/v2\/media?parent=427"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}