The Ceph Blog

Featured Post

v0.88 released

This is the first development release after Giant. The two main features merged this round are the new AsyncMessenger (an alternative implementation of the network layer) from Haomai Wang at UnitedStack, and support for POSIX file locks in ceph-fuse and libcephfs from Yan, Zheng. There is also a big pile of smaller items that re merged while we were stabilizing Giant, including a range of smaller performance and bug fixes and some new tracepoints for LTTNG.

NOTABLE CHANGES

  • ceph-disk: Scientific Linux support (Dan van der Ster)
  • ceph-disk: respect –statedir for keyring (Loic Dachary)
  • ceph-fuse, libcephfs: POSIX file lock support (Yan, Zheng)
  • ceph-fuse, libcephfs: fix cap flush overflow (Greg Farnum, Yan, Zheng)
  • ceph-fuse, libcephfs: fix root inode xattrs (Yan, Zheng)
  • ceph-fuse, libcephfs: preserve dir ordering (#9178 Yan, Zheng)
  • ceph-fuse, libcephfs: trim inodes before reconnecting to MDS (Yan, Zheng)
  • ceph: do not parse injectargs twice (Loic Dachary)
  • ceph: make ‘ceph -s’ output more readable (Sage Weil)
  • ceph: new ‘ceph tell mds.$name_or_rank_or_gid’ (John Spray)
  • ceph: test robustness (Joao Eduardo Luis)
  • ceph_objectstore_tool: behave with sharded flag (#9661 David Zafman)
  • cephfs-journal-tool: fix journal import (#10025 John Spray)
  • cephfs-journal-tool: skip up to expire_pos (#9977 John Spray)
  • cleanup rados.h definitions with macros (Ilya Dryomov)
  • common: shared_cache unit tests (Cheng Cheng)
  • config: add $cctid meta variable (Adam Crume)
  • crush: fix buffer overrun for poorly formed rules (#9492 Johnu George)
  • crush: improve constness (Loic Dachary)
  • crushtool: add –location <id> command (Sage Weil, Loic Dachary)
  • default to libnss instead of crypto++ (Federico Gimenez)
  • doc: ceph osd reweight vs crush weight (Laurent Guerby)
  • doc: document the LRC per-layer plugin configuration (Yuan Zhou)
  • doc: erasure code doc updates (Loic Dachary)
  • doc: misc updates (Alfredo Deza, VRan Liu)
  • doc: preflight doc fixes (John Wilkins)
  • doc: update PG count guide (Gerben Meijer, Laurent Guerby, Loic Dachary)
  • keyvaluestore: misc fixes (Haomai Wang)
  • keyvaluestore: performance improvements (Haomai Wang)
  • librados: add rados_pool_get_base_tier() call (Adam Crume)
  • librados: cap buffer length (Loic Dachary)
  • librados: fix objecter races (#9617 Josh Durgin)
  • libradosstriper: misc fixes (Sebastien Ponce)
  • librbd: add missing python docstrings (Jason Dillaman)
  • librbd: add readahead (Adam Crume)
  • librbd: fix cache tiers in list_children and snap_unprotect (Adam Crume)
  • librbd: fix performance regression in ObjectCacher (#9513 Adam Crume)
  • librbd: lttng tracepoints (Adam Crume)
  • librbd: misc fixes (Xinxin Shu, Jason Dillaman)
  • mds: fix sessionmap lifecycle bugs (Yan, Zheng)
  • mds: initialize root inode xattr version (Yan, Zheng)
  • mds: introduce auth caps (John Spray)
  • mds: misc bugs (Greg Farnum, John Spray, Yan, Zheng, Henry Change)
  • misc coverity fixes (Danny Al-Gaaf)
  • mon: add ‘ceph osd rename-bucket …’ command (Loic Dachary)
  • mon: clean up auth list output (Loic Dachary)
  • mon: fix ‘osd crush link’ id resolution (John Spray)
  • mon: fix misc error paths (Joao Eduardo Luis)
  • mon: fix paxos off-by-one corner case (#9301 Sage Weil)
  • mon: new ‘ceph pool ls [detail]’ command (Sage Weil)
  • mon: wait for writeable before cross-proposing (#9794 Joao Eduardo Luis)
  • msgr: avoid useless new/delete (Haomai Wang)
  • msgr: fix delay injection bug (#9910 Sage Weil, Greg Farnum)
  • msgr: new AsymcMessenger alternative implementation (Haomai Wang)
  • msgr: prefetch data when doing recv (Yehuda Sadeh)
  • osd: add erasure code corpus (Loic Dachary)
  • osd: add misc tests (Loic Dachary, Danny Al-Gaaf)
  • osd: cleanup boost optionals (William Kennington)
  • osd: expose non-journal backends via ceph-osd CLI (Hoamai Wang)
  • osd: fix JSON output for stray OSDs (Loic Dachary)
  • osd: fix ioprio options (Loic Dachary)
  • osd: fix transaction accounting (Jianpeng Ma)
  • osd: misc optimizations (Xinxin Shu, Zhiqiang Wang, Xinze Chi)
  • osd: use FIEMAP_FLAGS_SYNC instead of fsync (Jianpeng Ma)
  • rados: fix put of /dev/null (Loic Dachary)
  • rados: parse command-line arguments more strictly (#8983 Adam Crume)
  • rbd-fuse: fix memory leak (Adam Crume)
  • rbd-replay-many (Adam Crume)
  • rbd-replay: –anonymize flag to rbd-replay-prep (Adam Crume)
  • rbd: fix ‘rbd diff’ for non-existent objects (Adam Crume)
  • rbd: fix error when striping with format 1 (Sebastien Han)
  • rbd: fix export for image sizes over 2GB (Vicente Cheng)
  • rbd: use rolling average for rbd bench-write throughput (Jason Dillaman)
  • rgw: send explicit HTTP status string (Yehuda Sadeh)
  • rgw: set length for keystone token validation request (#7796 Yehuda Sadeh, Mark Kirkwood)
  • udev: fix rules for CentOS7/RHEL7 (Loic Dachary)
  • use clock_gettime instead of gettimeofday (Jianpeng Ma)
  • vstart.sh: set up environment for s3-tests (Luis Pabon)

GETTING CEPH

Earlier Posts

v0.87 Giant released

This release will form the basis for the stable release Giant, v0.87.x. Highlights for Giant include:

  • RADOS Performance: a range of improvements have been made in the OSD and client-side librados code that improve the throughput on flash backends and improve parallelism and scaling on fast machines.
  • CephFS: we have fixed a raft of bugs in CephFS and built some basic journal recovery and diagnostic tools. Stability and performance of single-MDS systems is vastly improved in Giant. Although we do not yet recommend CephFS for production deployments, we do encourage testing for non-critical workloads so that we can better guage the feature, usability, performance, and stability gaps.
  • Local Recovery Codes: the OSDs now support an erasure-coding scheme that stores some additional data blocks to reduce the IO required to recover from single OSD failures.
  • Degraded vs misplaced: the Ceph health reports from ‘ceph -s’ and related commands now make a distinction between data that is degraded (there are fewer than the desired number of copies) and data that is misplaced (stored in the wrong location in the cluster). The distinction is important because the latter does not compromise data safety.
  • Tiering improvements: we have made several improvements to the cache tiering implementation that improve performance. Most notably, objects are not promoted into the cache tier by a single read; they must be found to be sufficiently hot before that happens.
  • Monitor performance: the monitors now perform writes to the local data store asynchronously, improving overall responsiveness.
  • Recovery tools: the ceph_objectstore_tool is greatly expanded to allow manipulation of an individual OSDs data store for debugging and repair purposes. This is most heavily used by our QA infrastructure to exercise recovery code.

read more…

v0.80.7 Firefly released

This release fixes a few critical issues with v0.80.6, particularly with clusters running mixed versions.

We recommend that all v0.80.x Firefly users upgrade to this release.

For more detailed information, see the complete changelog.

NOTABLE CHANGES

  • osd: fix invalid memory reference in log trimming (#9731 Samuel Just)
  • osd: fix use-after-free in cache tiering code (#7588 Sage Weil)
  • osd: remove bad backfill assertion for mixed-version clusters (#9696 Samuel Just)

GETTING CEPH

A few days ago I made a challenge to the open storage community to support the Ada Initiative‘s work to improve gender diversity in open source and open data communities, and offered to match contributions if we can reach $8192.  I’m pleased to say that the response so far has been great, and we’re now over half way there!  I’d like to thank those who have contributed so far, including

Josef Bacik
Dave McAllister
Eric Sandeen
Zulah and Carlos
Geoff Arnold
Andreas Dilger
Tom Lyon
Greg Farnum
Travis Rhoden
Garrett D’Amore
Erin M. Evans
Travis Rhoden
zab
Dan V
cjs
Peter Tribble
Bryan Horstmann-Allen
Patrick McGarry
7 anonymous donors

I’m quite pleased to see Linux, Lustre, GlusterFS, and OpenZFS / Illumos represented on this list! It’s also great to see that this is an issue that the Illumos community has already identified and recently called out:

 

 

Increasing awareness of the issue and showing broad support for these campaigns is just as important as the money raised, so please contribute or help spread the word even if it is a token amount!

I would love to see some similar attention to this issue in the OpenStack Cinder and Swift communities. What do you think?

Donate now

v0.80.6 Firefly released

This is a major bugfix release for firefly, fixing a range of issues in the OSD and monitor, particularly with cache tiering. There are also important fixes in librados, with the watch/notify mechanism used by librbd, and in radosgw.

A few pieces of new functionality of been backported, including improved ‘ceph df’ output (view amount of writeable space per pool), support for non-default cluster names when using sysvinit or systemd, and improved (and fixed) support for dmcrypt.

We recommend that all v0.80.x Firefly users upgrade to this release.

For more detailed information, see the complete changelog.

NOTABLE CHANGES

read more…

I’d like to take a moment away from your regularly scheduled storage revolution to talk about the Ada Initiative: who they are, what they do, and why it is important to open source storage communities. I’m also going to challenge you to raise $8192 for them, and I’ll match that dollar for dollar if you do. (If you already know who they are and support their work, go ahead and donate now.)

Research, experience, and common sense have demonstrated that diverse communities perform better: they are more dynamic, they generate better ideas, and they are more pleasant to be a part of. On the gender axis, however, most open source communities–including storage–are extremely homogeneous, with far less than 10% participation from women. It surprised me to learn that this is significantly lower than software engineering in general, where you find women make up about 30% of the community.

This frustrates me on two levels.  First, open source is about building communities around code that bring together diverse organizations and interests in pursuit of a common goal.  Why is something as fundamental as being inclusive of women a problem here?  Our “open” communities should be setting the standard, not lagging behind when it comes to diversity.

Second, I am passionate about Ceph not because I think it is the end-all solution to everyone’s storage problems, but because it is playing an important role in breaking the stranglehold that proprietary vendors have over a large and increasingly critical industry.  If we are going to win the larger battle of making the industry-leading, state of the art, de facto choice for storage an open platform, we will need all the help we can get: the incumbents are well-entrenched, they are better funded, and (it turns out) they are doing a better job of attracting diverse talent.

The Ada Initiative is one of the few organizations who is addressing gender diversity head-on, with a specific focus on open source and open data communities.  You might already know about their successful campaign to get most open source conferences to adopt codes of conduct, which make women and other marginalized groups more likely to attend.  Their AdaCamp conferences for women in open tech/culture, and their Ally Skills Workshops which teach men how to support women in everyday ways have proven to be extremely popular and effective both in welcoming and empowering women.  These programs have proven so successful, in fact, that (for lack of staff and funding) they are currently unable to meet the full demand for them: all three AdaCamps this year sold out months early, and they are booking solid for Ally Skills Workshops for the next 3 months.

As Inktank and as DreamHost we were proud to be early supporters of the Ada Initiative.  Today, I am proud to continue that support with a personal challenge to the open source storage community:

If you raise $8192 by next Wednesday, I will match that contribution dollar for dollar.

This challenge applies to the larger open source storage world, including the Ceph, Gluster, Swift, and Linux storage and file systems communities, and goes until Wednesday, October 8th, when the Ada Initiative’s fundraising drive ends.

Donate now

Ceph Developer Summit: Hammer

As many of you Ceph Day attendees are no doubt aware, we’re fast approaching the release date for the ‘Giant’ release of Ceph. With that, it’s time to get together at another virtual Ceph Developer Summit and chat about what development work is going in to the ‘Hammer’ release. Blueprint submissions are open now, so if you have any work you would like to contribute or request of our community developers, please submit it as soon as possible to ensure it gets a CDS slot.

The rough schedule of CDS and Hammer in general should look something like this:

Date Milestone
30 SEP Blueprint submissions begin
17 OCT Blueprint submissions end
21 OCT Summit agenda announced
28 OCT Ceph Developer Summit: Day 1
29 OCT Ceph Developer Summit: Day 2 (if needed)
January 2015 Hammer Release

If there are enough sessions we are exploring the possibility of expanding our event into three days, but that will be predicated on the blueprint workload. As always, this event will be an online event (utilizing the BlueJeans system) so that everyone can attend from their own timezone. If you are interested in submitting a blueprint or collaborating on an existing blueprint, please click the big red button below!

 

Submit Blueprint

scuttlemonkey out

v0.67.11 Dumpling released

This stable update for Dumpling fixes several important bugs that affect a small set of users.

We recommend that all Dumpling users upgrade at their convenience. If none of these issues are affecting your deployment there is no urgency.

NOTABLE CHANGES

  • common: fix sending dup cluster log items (#9080 Sage Weil)
  • doc: several doc updates (Alfredo Deza)
  • libcephfs-java: fix build against older JNI headesr (Greg Farnum)
  • librados: fix crash in op timeout path (#9362 Matthias Kiefer, Sage Weil)
  • librbd: fix crash using clone of flattened image (#8845 Josh Durgin)
  • librbd: fix error path cleanup when failing to open image (#8912 Josh Durgin)
  • mon: fix crash when adjusting pg_num before any OSDs are added (#9052 Sage Weil)
  • mon: reduce log noise from paxos (Aanchal Agrawal, Sage Weil)
  • osd: allow scrub and snap trim thread pool IO priority to be adjusted (Sage Weil)
  • osd: fix mount/remount sync race (#9144 Sage Weil)

GETTING CEPH

v0.85 released

This is the second-to-last development release before Giant that contains new functionality. The big items to land during this cycle are the messenger refactoring from Matt Benjmain that lays some groundwork for RDMA support, a performance improvement series from SanDisk that improves performance on SSDs, lots of improvements to our new standalone civetweb-based RGW frontend, and a new ‘osd blocked-by’ mon command that allows admins to easily identify which OSDs are blocking peering progress. The other big change is that the OSDs and Monitors now distinguish between “misplaced” and “degraded” objects: the latter means there are fewer copies than we’d like, while the former simply means the are not stored in the locations where we want them to be.

Also of note is a change to librbd that enables client-side caching by default. This is coupled with another option that makes the cache write-through until a “flush” operations is observed: this implies that the librbd user (usually a VM guest OS) supports barriers and flush and that it is safe for the cache to switch into writeback mode without compromising data safety or integrity. It has long been recommended practice that these options be enabled (e.g., in OpenStack environments) but until now it has not been the default.

We have frozen the tree for the looming Giant release, and the next development release will be a release candidate with a final batch of new functionality.

read more…

v0.84 released

The next Ceph development release is here! This release contains several meaty items, including some MDS improvements for journaling, the ability to remove the CephFS file system (and name it), several mon cleanups with tiered pools, several OSD performance branches, a new “read forward” RADOS caching mode, a prototype Kinetic OSD backend, and various radosgw improvements (especially with the new standalone civetweb frontend). And there are a zillion OSD bug fixes. Things are looking pretty good for the Giant release that is coming up in the next month.

UPGRADING

  • The *_kb perf counters on the monitor have been removed. These are replaced with a new set of *_bytes counters (e.g., cluster_osd_kb is replaced by cluster_osd_bytes).
  • The rd_kb and wr_kb fields in the JSON dumps for pool stats (accessed via the ‘ceph df detail -f json-pretty’ and related commands) have been replaced with corresponding *_bytes fields. Similarly, the ‘total_space’, ‘total_used’, and ‘total_avail’ fields are replaced with ‘total_bytes’, ‘total_used_bytes’, and ‘total_avail_bytes’ fields.
  • The ‘rados df –format=json’ output ‘read_bytes’ and ‘write_bytes’ fields were incorrectly reporting ops; this is now fixed.
  • The ‘rados df –format=json’ output previously included ‘read_kb’ and ‘write_kb’ fields; these have been removed. Please use ‘read_bytes’ and ‘write_bytes’ instead (and divide by 1024 if appropriate).

read more…

Page 1 of 1412345...10...Last »
© 2014, Inktank Storage, Inc.. All rights reserved.