// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*- // vim: ts=8 sw=2 smarttab #ifndef CEPH_LIBRBD_CACHE_WRITE_AROUND_OBJECT_DISPATCH_H #define CEPH_LIBRBD_CACHE_WRITE_AROUND_OBJECT_DISPATCH_H #include "librbd/io/ObjectDispatchInterface.h" #include "include/interval_set.h" #include "common/ceph_mutex.h" #include "librbd/io/Types.h" #include #include #include struct Context; namespace librbd { struct ImageCtx; namespace cache { template class WriteAroundObjectDispatch : public io::ObjectDispatchInterface { public: static WriteAroundObjectDispatch* create(ImageCtxT* image_ctx, size_t max_dirty, bool writethrough_until_flush) { return new WriteAroundObjectDispatch(image_ctx, max_dirty, writethrough_until_flush); } WriteAroundObjectDispatch(ImageCtxT* image_ctx, size_t max_dirty, bool writethrough_until_flush); ~WriteAroundObjectDispatch() override; io::ObjectDispatchLayer get_dispatch_layer() const override { return io::OBJECT_DISPATCH_LAYER_CACHE; } void init(); void shut_down(Context* on_finish) override; bool read( uint64_t object_no, io::ReadExtents* extents, IOContext io_context, int op_flags, int read_flags, const ZTracer::Trace &parent_trace, uint64_t* version, int* object_dispatch_flags, io::DispatchResult* dispatch_result, Context** on_finish, Context* on_dispatched) override; bool discard( uint64_t object_no, uint64_t object_off, uint64_t object_len, IOContext io_context, int discard_flags, const ZTracer::Trace &parent_trace, int* object_dispatch_flags, uint64_t* journal_tid, io::DispatchResult* dispatch_result, Context**on_finish, Context* on_dispatched) override; bool write( uint64_t object_no, uint64_t object_off, ceph::bufferlist&& data, IOContext io_context, int op_flags, int write_flags, std::optional assert_version, const ZTracer::Trace &parent_trace, int* object_dispatch_flags, uint64_t* journal_tid, io::DispatchResult* dispatch_result, Context**on_finish, Context* on_dispatched) override; bool write_same( uint64_t object_no, uint64_t object_off, uint64_t object_len, io::LightweightBufferExtents&& buffer_extents, ceph::bufferlist&& data, IOContext io_context, int op_flags, const ZTracer::Trace &parent_trace, int* object_dispatch_flags, uint64_t* journal_tid, io::DispatchResult* dispatch_result, Context**on_finish, Context* on_dispatched) override; bool compare_and_write( uint64_t object_no, uint64_t object_off, ceph::bufferlist&& cmp_data, ceph::bufferlist&& write_data, IOContext io_context, int op_flags, const ZTracer::Trace &parent_trace, uint64_t* mismatch_offset, int* object_dispatch_flags, uint64_t* journal_tid, io::DispatchResult* dispatch_result, Context** on_finish, Context* on_dispatched) override; bool flush( io::FlushSource flush_source, const ZTracer::Trace &parent_trace, uint64_t* journal_tid, io::DispatchResult* dispatch_result, Context** on_finish, Context* on_dispatched) override; bool list_snaps( uint64_t object_no, io::Extents&& extents, io::SnapIds&& snap_ids, int list_snap_flags, const ZTracer::Trace &parent_trace, io::SnapshotDelta* snapshot_delta, int* object_dispatch_flags, io::DispatchResult* dispatch_result, Context** on_finish, Context* on_dispatched) override { return false; } bool invalidate_cache(Context* on_finish) override { return false; } bool reset_existence_cache(Context* on_finish) override { return false; } void extent_overwritten( uint64_t object_no, uint64_t object_off, uint64_t object_len, uint64_t journal_tid, uint64_t new_journal_tid) override { } int prepare_copyup( uint64_t object_no, io::SnapshotSparseBufferlist* snapshot_sparse_bufferlist) override { return 0; } private: struct QueuedIO { QueuedIO(uint64_t length, Context* on_finish, Context* on_dispatched) : length(length), on_finish(on_finish), on_dispatched(on_dispatched) { } uint64_t length; Context* on_finish; Context* on_dispatched; }; struct QueuedFlush { QueuedFlush(Context* on_finish, Context* on_dispatched) : on_finish(on_finish), on_dispatched(on_dispatched) { } Context* on_finish; Context* on_dispatched; }; struct BlockedIO : public QueuedIO { BlockedIO(uint64_t offset, uint64_t length, Context* on_finish, Context* on_dispatched) : QueuedIO(length, on_finish, on_dispatched), offset(offset) { } uint64_t offset; }; typedef std::map QueuedIOs; typedef std::map QueuedFlushes; typedef std::map BlockedObjectIOs; typedef std::map BlockedIOs; typedef std::map Contexts; typedef std::set Tids; typedef interval_set InFlightObjectExtents; typedef std::map InFlightExtents; ImageCtxT* m_image_ctx; size_t m_init_max_dirty; size_t m_max_dirty; ceph::mutex m_lock; bool m_user_flushed = false; uint64_t m_last_tid = 0; uint64_t m_in_flight_bytes = 0; Tids m_in_flight_io_tids; InFlightExtents m_in_flight_extents; BlockedIOs m_blocked_ios; QueuedIOs m_queued_ios; Tids m_queued_or_blocked_io_tids; BlockedIOs m_blocked_unoptimized_ios; QueuedFlushes m_queued_flushes; Contexts m_in_flight_flushes; Contexts m_pending_flushes; int m_pending_flush_error = 0; bool dispatch_unoptimized_io(uint64_t object_no, uint64_t object_off, uint64_t object_len, io::DispatchResult* dispatch_result, Context* on_dispatched); bool dispatch_io(uint64_t object_no, uint64_t object_off, uint64_t object_len, int op_flags, io::DispatchResult* dispatch_result, Context** on_finish, Context* on_dispatch); bool block_overlapping_io(InFlightObjectExtents* in_flight_object_extents, uint64_t object_off, uint64_t object_len); void unblock_overlapping_ios(uint64_t object_no, uint64_t object_off, uint64_t object_len, Contexts* unoptimized_io_dispatches); bool can_dispatch_io(uint64_t tid, uint64_t length); void handle_in_flight_io_complete(int r, uint64_t tid, uint64_t object_no, uint64_t object_off, uint64_t object_len); void handle_in_flight_flush_complete(int r, uint64_t tid); QueuedIOs collect_ready_ios(); Contexts collect_ready_flushes(); Contexts collect_finished_flushes(); }; } // namespace cache } // namespace librbd extern template class librbd::cache::WriteAroundObjectDispatch; #endif // CEPH_LIBRBD_CACHE_WRITE_AROUND_OBJECT_DISPATCH_H