Hello Cloud people,
Finally, with a lot of help from the upstream OpenNebula developers we have
finished packaging this for Fedora 17.
Due to the urgency of Fedora 17 now being beta, I would really appreciate if
people can review the package.
Please let me know you can help review.
Talking with some Fedora's guys in Openstack Design and Summit at SF, I
decide write this my first post and share some work around OpenStack Essex
specs. The specs are hosted in github.com. I think that could have
something to do but, the keystone, glance and nova are working properly.
Suggestions are appreciated.
we apologize if you receive multiple copies of this CFP.
CALL FOR PAPERS
7th Workshop on
Virtualization in High-Performance Cloud Computing
as part of Euro-Par 2012, Rhodes Island, Greece
Date: August 28, 2012
Workshop URL: http://vhpc.org
Rolling abstract submission
June 4, 2012 - Full paper submission
Virtualization has become a common abstraction layer in modern
data centers, enabling resource owners to manage complex
infrastructure independently of their applications. Conjointly,
virtualization is becoming a driving technology for a manifold of
industry grade IT services. The cloud concept includes the notion
of a separation between resource owners and users, adding services
such as hosted application frameworks and queueing. Utilizing the
same infrastructure, clouds carry significant potential for use in
high-performance scientific computing. The ability of clouds to provide
for requests and releases of vast computing resources dynamically and
close to the marginal cost of providing the services is unprecedented in
the history of scientific and commercial computing.
Distributed computing concepts that leverage federated resource
access are popular within the grid community, but have not seen
previously desired deployed levels so far. Also, many of the scientific
data centers have not adopted virtualization or cloud concepts yet.
This workshop aims to bring together industrial providers with the
scientific community in order to foster discussion, collaboration
and mutual exchange of knowledge and experience.
The workshop will be one day in length, composed of 20 min
paper presentations, each followed by 10 min discussion sections.
Presentations may be accompanied by interactive demonstrations.
Topics of interest include, but are not limited to:
Higher-level cloud architectures, focusing on issues such as:
- Languages for describing highly-distributed compute jobs
- Workload characterization for VM-based environments
- Optimized communication libraries/protocols in the cloud
- Cross-layer optimization of numeric algorithms on VM infrastructure
- System and process/bytecode VM convergence
- Cloud frameworks and API sets
- Checkpointing/migration of large compute jobs
- Instrumentation interfaces and languages
- VMM performance (auto-)tuning on various load types
- Cloud reliability, fault-tolerance, and security
- Software as a Service (SaaS) architectures
- Research and education use cases
- Virtualization in cloud, cluster and grid environments
- Cross-layer VM optimizations
- Cloud use cases including optimizations
- VM-based cloud performance modelling
- Performance and cost modelling
Lower-level design challenges for Hypervisors, VM-aware I/O devices,
hardware accelerators or filesystems in VM environments, especially:
- Cloud, grid and distributed filesystems
- Hardware for I/O virtualization (storage/network/accelerators)
- Storage and network I/O subsystems in virtualized environments
- Novel software approaches to I/O virtualization
- Paravirtualized I/O subsystems for modified/unmodified guests
- Virtualization-aware cluster interconnects
- Direct device assignment
- NUMA-aware subsystems in virtualized environments
- Hardware Accelerators in virtualization (GPUs/FPGAs)
- Hardware extensions for virtualization
- VMMs/Hypervisors for embedded systems
Data Center management methods, including:
- QoS and and service levels
- VM cloud and cluster distribution algorithms
- VM load-balancing in Clouds
- Hypervisor extensions and tools for cluster and grid computing
- Fault tolerant VM environments
- Virtual machine monitor platforms
- Management, deployment and monitoring of VM-based environments
- Cluster provisioning in the Cloud
Papers submitted to the workshop will be reviewed by at least two
members of the program committee and external reviewers. Submissions
should include abstract, key words, the e-mail address of the
corresponding author, and must not exceed 10 pages, including tables
and figures at a main font size no smaller than 11 point. Submission
of a paper should be regarded as a commitment that, should the paper
be accepted, at least one of the authors will register and attend the
conference to present the work.
Accepted papers will be published in the Springer LNCS series - the
format must be according to the Springer LNCS Style. Initial
submissions are in PDF; authors of accepted papers will be requested
to provide source files.
Format Guidelines: http://www.springer.de/comp/lncs/authors.html
Abstract Submission Link: http://edas.info/newPaper.php?c=11943
Rolling abstract submission
June 4, 2012 - Full paper submission
June 29, 2012 - Acceptance notification
July 20, 2012 - Camera-ready version due
August 28, 2012 - Workshop Date
Michael Alexander (chair), TU Wien, Austria
Gianluigi Zanetti (co-chair), CRS4, Italy
Anastassios Nanos (co-chair), NTUA, Greece
Paolo Anedda, CRS4, Italy
Giovanni Busonera, CRS4, Italy
Brad Calder, Microsoft, USA
Roberto Canonico, University of Napoli Federico II, Italy
Tommaso Cucinotta, Alcatel-Lucent Bell Labs, Ireland
Werner Fischer, Thomas-Krenn AG, Germany
William Gardner, University of Guelph, USA
Marcus Hardt, Forschungszentrum Karlsruhe, Germany
Sverre Jarp, CERN, Switzerland
Shantenu Jha, Louisiana State University, USA
Xuxian Jiang, NC State, USA
Nectarios Koziris, National Technical University of Athens, Greece
Simone Leo, CRS4, Italy
Ignacio Llorente, Universidad Complutense de Madrid, Spain
Naoya Maruyama, Tokyo Institute of Technology, Japan
Jean-Marc Menaud, Ecole des Mines de Nantes, France
Dimitrios Nikolopoulos, Foundation for Research&Technology Hellas, Greece
Jose Renato Santos, HP Labs, USA
Walter Schwaiger, TU Wien, Austria
Yoshio Turner, HP Labs, USA
Kurt Tutschku, University of Vienna, Austria
Lizhe Wang, Indiana University, USA
Chao-Tung Yang, Tunghai University, Taiwan
DURATION: Workshop Duration is one day.
The workshop will be held as part of Euro-Par 2012.
Euro-Par 2012: http://europar2012.cti.gr/
-----BEGIN PGP SIGNED MESSAGE-----
Here are the F17 Beta AMIS I did a lot of cutting and pasting to get
them here and it seems that while I can time travel my tardis doesnt
let me go back and fix typos. So if i managed to mess one up please let
me know, Thanks.
-----BEGIN PGP SIGNATURE-----
Version: GnuPG v2.0.18 (GNU/Linux)
-----END PGP SIGNATURE-----
Because Ruth is awesome (HI RUTH!) she took a swag at putting together
some of the input from here:
And turned it into this:
I think it is a 5x7 postcard-style/weight piece.
It's still pretty drafty but seeking comments/edits (a lot of this was
My main ... not concern, but, well.. we're in this awkward spot where
things are done for the Essex release, but F17 itself isn't actually
out, so I'm not sure if it's appropriate to pull in Quantum, Keystone,
Horizon into the top set of bulletpoints. (That said, I haven't looked
to see if Essex is also available now for F16, which might make this
less of a conundrum.)
TAKING FEEDBACK (and praise for ruth). (And yes, if we get it done
between now and Monday, they can be there on Wednesday for boothiness on
Thursday/Friday during the conference part.)
On Apr 9, 2012 3:26 PM, "Alan Gutierrez" <alan(a)prettyrobots.com> wrote:
> Thank you for responding. That is what I expected to find, the server's
ssh key printed to the console output. But, I don't see it there. Not when
using one of the Fedora 16 AMIs.
> Here's a gist of the output of the California hosted Fedora 16 i386 AMI
> I'm using a small instance.
> It this something that is a known issue? Or do people generally build
their own instances using BoxGrinder?
I suspect cloud-init prints the fingerprints and then systemd eats the
output as it blindly redirects all daemons' output to syslog. That can
probably be fixed in cloud-init's unit files if that is the case.