/* * Copyright (C) 2006-2014 David Robillard * Copyright (C) 2007-2012 Carl Hetherington * Copyright (C) 2007-2017 Paul Davis * Copyright (C) 2007 Doug McLain * Copyright (C) 2014-2015 Ben Loftis * Copyright (C) 2014-2017 Nick Mainsbridge * Copyright (C) 2014-2019 Robin Gareus * Copyright (C) 2015-2016 Tim Mayberry * * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation; either version 2 of the License, or * (at your option) any later version. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License along * with this program; if not, write to the Free Software Foundation, Inc., * 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. */ #include #include #include #include #include #include #include "ardour/playlist.h" #include "ardour/audioregion.h" #include "ardour/audiosource.h" #include "ardour/profile.h" #include "ardour/region_fx_plugin.h" #include "ardour/session.h" #include "pbd/memento_command.h" #include "evoral/Curve.h" #include "gtkmm2ext/gtk_ui.h" #include "gtkmm2ext/utils.h" #include "gtkmm2ext/colors.h" #include "canvas/rectangle.h" #include "canvas/polygon.h" #include "canvas/poly_line.h" #include "canvas/line.h" #include "canvas/text.h" #include "canvas/xfade_curve.h" #include "canvas/debug.h" #include "waveview/debug.h" #include "streamview.h" #include "audio_region_view.h" #include "audio_time_axis.h" #include "enums_convert.h" #include "public_editor.h" #include "audio_region_editor.h" #include "audio_streamview.h" #include "mergeable_line.h" #include "region_gain_line.h" #include "control_point.h" #include "paste_context.h" #include "ghostregion.h" #include "audio_time_axis.h" #include "rgb_macros.h" #include "gui_thread.h" #include "ui_config.h" #include "pbd/i18n.h" using namespace std; using namespace ARDOUR; using namespace PBD; using namespace Editing; using namespace ArdourCanvas; static double const handle_size = 10; /* height of fade handles */ Cairo::RefPtr AudioRegionView::pending_peak_pattern; static Cairo::RefPtr create_pending_peak_pattern() { cairo_surface_t * is = cairo_image_surface_create (CAIRO_FORMAT_ARGB32, 8, 8); // create checker pattern unsigned char *img = cairo_image_surface_get_data (is); cairo_surface_flush (is); const int stride = cairo_image_surface_get_stride (is); for (int y = 0; y < 8; ++y) { for (int x = 0; x < 8; ++x) { const int off = (y * stride + x * 4); uint32_t *pixel = (uint32_t*) &img[off]; if ((x < 4) ^ (y < 4)) { *pixel = 0xa0000000; } else { *pixel = 0x40000000; } } } cairo_surface_mark_dirty (is); cairo_pattern_t* pat = cairo_pattern_create_for_surface (is); cairo_pattern_set_extend (pat, CAIRO_EXTEND_REPEAT); Cairo::RefPtr p (new Cairo::Pattern (pat, false)); cairo_surface_destroy (is); return p; } AudioRegionView::AudioRegionView (ArdourCanvas::Container *parent, RouteTimeAxisView &tv, std::shared_ptr r, double spu, uint32_t basic_color) : RegionView (parent, tv, r, spu, basic_color) , fade_in_handle(0) , fade_out_handle(0) , fade_in_trim_handle(0) , fade_out_trim_handle(0) , pending_peak_data(0) , start_xfade_curve (0) , start_xfade_rect (0) , _start_xfade_visible (false) , end_xfade_curve (0) , end_xfade_rect (0) , _end_xfade_visible (false) , _amplitude_above_axis(1.0) , trim_fade_in_drag_active(false) , trim_fade_out_drag_active(false) , _rfx_id (0) , _rdx_param (UINT32_MAX) , _ignore_line_change (false) { } AudioRegionView::AudioRegionView (ArdourCanvas::Container *parent, RouteTimeAxisView &tv, std::shared_ptr r, double spu, uint32_t basic_color, bool recording, TimeAxisViewItem::Visibility visibility) : RegionView (parent, tv, r, spu, basic_color, recording, visibility) , fade_in_handle(0) , fade_out_handle(0) , fade_in_trim_handle(0) , fade_out_trim_handle(0) , pending_peak_data(0) , start_xfade_curve (0) , start_xfade_rect (0) , _start_xfade_visible (false) , end_xfade_curve (0) , end_xfade_rect (0) , _end_xfade_visible (false) , _amplitude_above_axis(1.0) , trim_fade_in_drag_active(false) , trim_fade_out_drag_active(false) , _rfx_id (0) , _rdx_param (UINT32_MAX) , _ignore_line_change (false) { } AudioRegionView::AudioRegionView (const AudioRegionView& other, std::shared_ptr other_region) : RegionView (other, std::shared_ptr (other_region)) , fade_in_handle(0) , fade_out_handle(0) , fade_in_trim_handle(0) , fade_out_trim_handle(0) , pending_peak_data(0) , start_xfade_curve (0) , start_xfade_rect (0) , _start_xfade_visible (false) , end_xfade_curve (0) , end_xfade_rect (0) , _end_xfade_visible (false) , _amplitude_above_axis (other._amplitude_above_axis) , trim_fade_in_drag_active(false) , trim_fade_out_drag_active(false) , _rfx_id (0) , _rdx_param (UINT32_MAX) , _ignore_line_change (false) { init (true); } void AudioRegionView::init (bool wfd) { // FIXME: Some redundancy here with RegionView::init. Need to figure out // where order is important and where it isn't... if (!pending_peak_pattern) { pending_peak_pattern = create_pending_peak_pattern(); } // needs to be created first, RegionView::init() calls set_height() pending_peak_data = new ArdourCanvas::Rectangle (group); CANVAS_DEBUG_NAME (pending_peak_data, string_compose ("pending peak rectangle for %1", region()->name())); pending_peak_data->set_outline_color (Gtkmm2ext::rgba_to_color (0, 0, 0, 0.0)); pending_peak_data->set_pattern (pending_peak_pattern); pending_peak_data->set_data ("regionview", this); pending_peak_data->hide (); RegionView::init (wfd); _amplitude_above_axis = 1.0; create_waves (); if (!_recregion) { fade_in_handle = new ArdourCanvas::Rectangle (group); CANVAS_DEBUG_NAME (fade_in_handle, string_compose ("fade in handle for %1", region()->name())); fade_in_handle->set_outline_color (Gtkmm2ext::rgba_to_color (0, 0, 0, 1.0)); fade_in_handle->set_fill_color (UIConfiguration::instance().color ("inactive fade handle")); fade_in_handle->set_data ("regionview", this); fade_in_handle->hide (); fade_out_handle = new ArdourCanvas::Rectangle (group); CANVAS_DEBUG_NAME (fade_out_handle, string_compose ("fade out handle for %1", region()->name())); fade_out_handle->set_outline_color (Gtkmm2ext::rgba_to_color (0, 0, 0, 1.0)); fade_out_handle->set_fill_color (UIConfiguration::instance().color ("inactive fade handle")); fade_out_handle->set_data ("regionview", this); fade_out_handle->hide (); fade_in_trim_handle = new ArdourCanvas::Rectangle (group); CANVAS_DEBUG_NAME (fade_in_handle, string_compose ("fade in trim handle for %1", region()->name())); fade_in_trim_handle->set_outline_color (Gtkmm2ext::rgba_to_color (0, 0, 0, 1.0)); fade_in_trim_handle->set_fill_color (UIConfiguration::instance().color ("inactive fade handle")); fade_in_trim_handle->set_data ("regionview", this); fade_in_trim_handle->hide (); fade_out_trim_handle = new ArdourCanvas::Rectangle (group); CANVAS_DEBUG_NAME (fade_out_handle, string_compose ("fade out trim handle for %1", region()->name())); fade_out_trim_handle->set_outline_color (Gtkmm2ext::rgba_to_color (0, 0, 0, 1.0)); fade_out_trim_handle->set_fill_color (UIConfiguration::instance().color ("inactive fade handle")); fade_out_trim_handle->set_data ("regionview", this); fade_out_trim_handle->hide (); } setup_fade_handle_positions (); if (!trackview.session()->config.get_show_region_fades()) { set_fade_visibility (false); } set_region_gain_line (); /* streamview will call set_height() */ //set_height (trackview.current_height()); // XXX not correct for Layered mode, but set_height() will fix later. region_muted (); region_sync_changed (); region_resized (ARDOUR::bounds_change); /* region_resized sets ghost region duration */ /* region_locked is a synonym for region_renamed () which is called in region_muted() above */ //region_locked (); envelope_active_changed (); fade_in_active_changed (); fade_out_active_changed (); reset_width_dependent_items (_pixel_width); if (fade_in_handle) { fade_in_handle->Event.connect (sigc::bind (sigc::mem_fun (PublicEditor::instance(), &PublicEditor::canvas_fade_in_handle_event), fade_in_handle, this, false)); } if (fade_out_handle) { fade_out_handle->Event.connect (sigc::bind (sigc::mem_fun (PublicEditor::instance(), &PublicEditor::canvas_fade_out_handle_event), fade_out_handle, this, false)); } if (fade_in_trim_handle) { fade_in_trim_handle->Event.connect (sigc::bind (sigc::mem_fun (PublicEditor::instance(), &PublicEditor::canvas_fade_in_handle_event), fade_in_trim_handle, this, true)); } if (fade_out_trim_handle) { fade_out_trim_handle->Event.connect (sigc::bind (sigc::mem_fun (PublicEditor::instance(), &PublicEditor::canvas_fade_out_handle_event), fade_out_trim_handle, this, true)); } set_colors (); setup_waveform_visibility (); get_canvas_frame()->set_data ("linemerger", (LineMerger*) this); _fx_line->canvas_group().raise_to_top (); /* XXX sync mark drag? */ } AudioRegionView::~AudioRegionView () { in_destructor = true; RegionViewGoingAway (this); /* EMIT_SIGNAL */ for (vector::iterator i = _data_ready_connections.begin(); i != _data_ready_connections.end(); ++i) { delete *i; } _data_ready_connections.clear (); for (list >::iterator i = feature_lines.begin(); i != feature_lines.end(); ++i) { delete ((*i).second); } /* all waveviews etc will be destroyed when the group is destroyed */ } std::shared_ptr AudioRegionView::audio_region() const { // "Guaranteed" to succeed... return std::dynamic_pointer_cast(_region); } void AudioRegionView::region_changed (const PropertyChange& what_changed) { ENSURE_GUI_THREAD (*this, &AudioRegionView::region_changed, what_changed); RegionView::region_changed (what_changed); if (what_changed.contains (ARDOUR::Properties::scale_amplitude)) { region_scale_amplitude_changed (); } if (what_changed.contains (ARDOUR::Properties::fade_in)) { fade_in_changed (); } if (what_changed.contains (ARDOUR::Properties::fade_out)) { fade_out_changed (); } if (what_changed.contains (ARDOUR::Properties::fade_in_active)) { fade_in_active_changed (); } if (what_changed.contains (ARDOUR::Properties::fade_out_active)) { fade_out_active_changed (); } if (what_changed.contains (ARDOUR::Properties::envelope_active)) { envelope_active_changed (); } if (what_changed.contains (ARDOUR::Properties::valid_transients)) { transients_changed (); } } void AudioRegionView::fade_in_changed () { reset_fade_in_shape (); } void AudioRegionView::fade_out_changed () { reset_fade_out_shape (); } void AudioRegionView::fade_in_active_changed () { if (start_xfade_rect) { if (audio_region()->fade_in_active()) { start_xfade_rect->set_fill (false); } else { start_xfade_rect->set_fill_color (UIConfiguration::instance().color_mod ("inactive crossfade", "inactive crossfade")); start_xfade_rect->set_fill (true); } } } void AudioRegionView::fade_out_active_changed () { if (end_xfade_rect) { if (audio_region()->fade_out_active()) { end_xfade_rect->set_fill (false); } else { end_xfade_rect->set_fill_color (UIConfiguration::instance().color_mod ("inactive crossfade", "inactive crossfade")); end_xfade_rect->set_fill (true); } } } uint32_t AudioRegionView::get_fill_color () const { Gtkmm2ext::Color c; const bool opaque = _region->opaque() || trackview.layer_display () == Stacked; if (_selected) { c = UIConfiguration::instance().color ("selected region base"); } else if (_recregion) { return UIConfiguration::instance().color ("recording rect"); } else if (!UIConfiguration::instance().get_color_regions_using_track_color()) { c = UIConfiguration::instance().color (fill_color_name); } else { c = fill_color; } if (opaque && ( !_dragging && !_region->muted () )) { return c; } else if (_dragging) { return UIConfiguration::instance().color_mod (c, "dragging region"); } else { return Gtkmm2ext::HSV(c).mod (UIConfiguration::instance().modifier ("transparent region base")).color (); } } void AudioRegionView::region_scale_amplitude_changed () { for (uint32_t n = 0; n < waves.size(); ++n) { waves[n]->gain_changed (); } region_renamed (); } void AudioRegionView::region_renamed () { std::string str = RegionView::make_name (); if (audio_region()->speed_mismatch (trackview.session()->sample_rate())) { str = string ("*") + str; } std::shared_ptr ar (audio_region()); const gain_t scale_amplitude = ar->scale_amplitude(); if (scale_amplitude != 1.0) { char tmp[32]; snprintf (tmp, 32, " (%.1fdB%s)", accurate_coefficient_to_dB (fabsf (scale_amplitude)), scale_amplitude < 0 ? " Ø" : ""); str += tmp; } set_item_name (str, this); set_name_text (str); } void AudioRegionView::region_resized (const PropertyChange& what_changed) { AudioGhostRegion* agr; RegionView::region_resized(what_changed); PropertyChange interesting_stuff; interesting_stuff.add (ARDOUR::Properties::start); interesting_stuff.add (ARDOUR::Properties::length); if (what_changed.contains (interesting_stuff)) { for (uint32_t n = 0; n < waves.size(); ++n) { waves[n]->region_resized (); } for (vector::iterator i = ghosts.begin(); i != ghosts.end(); ++i) { if ((agr = dynamic_cast(*i)) != 0) { for (vector::iterator w = agr->waves.begin(); w != agr->waves.end(); ++w) { (*w)->region_resized (); } } } /* hide transient lines that extend beyond the region */ list >::iterator l; samplepos_t first = _region->first_sample(); samplepos_t last = _region->last_sample(); for (l = feature_lines.begin(); l != feature_lines.end(); ++l) { if (l->first < first || l->first >= last) { l->second->hide(); } else { l->second->show(); } } } } void AudioRegionView::reset_width_dependent_items (double pixel_width) { if (pixel_width == _width) { return; } RegionView::reset_width_dependent_items(pixel_width); assert(_pixel_width == pixel_width); pending_peak_data->set_x1(pixel_width); if (pixel_width <= 20.0 || _height < 5.0 || !trackview.session()->config.get_show_region_fades()) { if (fade_in_handle) { fade_in_handle->hide(); } if (fade_out_handle) { fade_out_handle->hide(); } if (fade_in_trim_handle) { fade_in_trim_handle->hide(); } if (fade_out_trim_handle) { fade_out_trim_handle->hide(); } if (start_xfade_rect) { start_xfade_rect->set_outline (false); } if (end_xfade_rect) { end_xfade_rect->set_outline (false); } } reset_fade_shapes (); /* Update feature lines */ AnalysisFeatureList analysis_features; _region->transients (analysis_features); if (feature_lines.size () != analysis_features.size ()) { cerr << "postponed freature line update.\n"; // XXX // AudioRegionView::transients_changed () will pick up on this return; } samplepos_t position = _region->position_sample(); AnalysisFeatureList::const_iterator i; list >::iterator l; double y1; if (_height >= NAME_HIGHLIGHT_THRESH) { y1 = _height - TimeAxisViewItem::NAME_HIGHLIGHT_SIZE - 1; } else { y1 = _height - 1; } for (i = analysis_features.begin(), l = feature_lines.begin(); i != analysis_features.end() && l != feature_lines.end(); ++i, ++l) { float x_pos = trackview.editor().sample_to_pixel ((*i) - position); (*l).first = *i; (*l).second->set (ArdourCanvas::Duple (x_pos, 2.0), ArdourCanvas::Duple (x_pos, y1)); } } void AudioRegionView::region_muted () { RegionView::region_muted(); set_waveform_colors (); } void AudioRegionView::setup_fade_handle_positions() { /* position of fade handle offset from the top of the region view */ double const handle_pos = 0.0; if (fade_in_handle) { fade_in_handle->set_y0 (handle_pos); fade_in_handle->set_y1 (handle_pos + handle_size); } if (fade_out_handle) { fade_out_handle->set_y0 (handle_pos); fade_out_handle->set_y1 (handle_pos + handle_size); } if (fade_in_trim_handle) { fade_in_trim_handle->set_y0 (_height - handle_size); fade_in_trim_handle->set_y1 (_height); } if (fade_out_trim_handle) { fade_out_trim_handle->set_y0 (_height - handle_size ); fade_out_trim_handle->set_y1 (_height); } } void AudioRegionView::set_height (gdouble height) { uint32_t gap = UIConfiguration::instance().get_vertical_region_gap (); float ui_scale = UIConfiguration::instance().get_ui_scale (); if (gap > 0 && ui_scale > 0) { gap = ceil (gap * ui_scale); } height = std::max (3.0, height - gap); if (height == _height) { return; } RegionView::set_height (height); pending_peak_data->set_y1 (height); RouteTimeAxisView& atv (*(dynamic_cast(&trackview))); // ick uint32_t nchans = atv.track()->n_channels().n_audio(); if (!tmp_waves.empty () || !waves.empty ()) { gdouble ht; if (!UIConfiguration::instance().get_show_name_highlight() || (height < NAME_HIGHLIGHT_THRESH)) { ht = height / (double) nchans; } else { ht = (height - NAME_HIGHLIGHT_SIZE) / (double) nchans; } uint32_t wcnt = waves.size(); for (uint32_t n = 0; n < wcnt; ++n) { gdouble yoff = floor (ht * n); waves[n]->set_height (ht); waves[n]->set_y_position (yoff); } wcnt = tmp_waves.size(); for (uint32_t n = 0; n < wcnt; ++n) { if (!tmp_waves[n]) { continue; } gdouble yoff = floor (ht * n); tmp_waves[n]->set_height (ht); tmp_waves[n]->set_y_position (yoff); } } if (_fx_line) { if ((height / nchans) < NAME_HIGHLIGHT_THRESH) { _fx_line->hide (); } else { update_envelope_visibility (); } _fx_line->set_height ((uint32_t) rint (height - NAME_HIGHLIGHT_SIZE) - 2); } reset_fade_shapes (); /* Update heights for any feature lines */ samplepos_t position = _region->position_sample(); list >::iterator l; double y1; if (_height >= NAME_HIGHLIGHT_THRESH) { y1 = _height - TimeAxisViewItem::NAME_HIGHLIGHT_SIZE - 1; } else { y1 = _height - 1; } for (l = feature_lines.begin(); l != feature_lines.end(); ++l) { float pos_x = trackview.editor().sample_to_pixel((*l).first - position); (*l).second->set (ArdourCanvas::Duple (pos_x, 2.0), ArdourCanvas::Duple (pos_x, y1)); } if (name_text) { name_text->raise_to_top(); } setup_fade_handle_positions(); } void AudioRegionView::reset_fade_shapes () { if (!trim_fade_in_drag_active) { reset_fade_in_shape (); } if (!trim_fade_out_drag_active) { reset_fade_out_shape (); } } void AudioRegionView::reset_fade_in_shape () { reset_fade_in_shape_width (audio_region(), audio_region()->fade_in()->back()->when.samples()); } void AudioRegionView::reset_fade_in_shape_width (std::shared_ptr ar, samplecnt_t width, bool drag_active) { trim_fade_in_drag_active = drag_active; if (fade_in_handle == 0) { return; } /* smallest size for a fade is 64 samples */ width = std::max ((samplecnt_t) 64, width); /* round here to prevent little visual glitches with sub-pixel placement */ double const pwidth = (double) width / samples_per_pixel; double const handle_left = pwidth; /* Put the fade in handle so that its left side is at the end-of-fade line */ fade_in_handle->set_x0 (handle_left); fade_in_handle->set_x1 (handle_left + handle_size); if (fade_in_trim_handle) { fade_in_trim_handle->set_x0 (0); fade_in_trim_handle->set_x1 (handle_size); } if (fade_in_handle->visible()) { //see comment for drag_start entered(); } if (pwidth < 5.f) { hide_start_xfade(); return; } if (!trackview.session()->config.get_show_region_fades()) { hide_start_xfade (); return; } double effective_height; if (_height >= NAME_HIGHLIGHT_THRESH) { effective_height = _height - NAME_HIGHLIGHT_SIZE; } else { effective_height = _height; } /* points *MUST* be in anti-clockwise order */ Points points; Points::size_type pi; std::shared_ptr list (audio_region()->fade_in()); Evoral::ControlList::const_iterator x; samplecnt_t length = list->length().samples(); points.assign (list->size(), Duple()); for (x = list->begin(), pi = 0; x != list->end(); ++x, ++pi) { const double p = (*x)->when.samples(); points[pi].x = (p * pwidth) / length; points[pi].y = effective_height - ((*x)->value * (effective_height - 1.)); } /* draw the line */ redraw_start_xfade_to (ar, width, points, effective_height, handle_left); /* ensure trim handle stays on top */ if (frame_handle_start) { frame_handle_start->raise_to_top(); } } void AudioRegionView::reset_fade_out_shape () { reset_fade_out_shape_width (audio_region(), audio_region()->fade_out()->back()->when.samples()); } void AudioRegionView::reset_fade_out_shape_width (std::shared_ptr ar, samplecnt_t width, bool drag_active) { trim_fade_out_drag_active = drag_active; if (fade_out_handle == 0) { return; } /* smallest size for a fade is 64 samples */ width = std::max ((samplecnt_t) 64, width); double const pwidth = floor(trackview.editor().sample_to_pixel (width)); /* the right edge should be right on the region frame is the pixel * width is zero. Hence the additional + 1.0 at the end. */ double const handle_right = rint(trackview.editor().sample_to_pixel (_region->length_samples()) - pwidth); double const trim_handle_right = rint(trackview.editor().sample_to_pixel (_region->length_samples())); /* Put the fade out handle so that its right side is at the end-of-fade line; */ fade_out_handle->set_x0 (handle_right - handle_size); fade_out_handle->set_x1 (handle_right); if (fade_out_trim_handle) { fade_out_trim_handle->set_x0 (1 + trim_handle_right - handle_size); fade_out_trim_handle->set_x1 (1 + trim_handle_right); } if (fade_out_handle->visible()) { //see comment for drag_start entered(); } /* don't show shape if its too small */ if (pwidth < 5) { hide_end_xfade(); return; } if (!trackview.session()->config.get_show_region_fades()) { hide_end_xfade(); return; } double effective_height; effective_height = _height; if (UIConfiguration::instance().get_show_name_highlight() && effective_height >= NAME_HIGHLIGHT_THRESH) { effective_height -= NAME_HIGHLIGHT_SIZE; } /* points *MUST* be in anti-clockwise order */ Points points; Points::size_type pi; std::shared_ptr list (audio_region()->fade_out()); Evoral::ControlList::const_iterator x; double length = list->length().samples(); points.assign (list->size(), Duple()); for (x = list->begin(), pi = 0; x != list->end(); ++x, ++pi) { const double p = (*x)->when.samples(); points[pi].x = _pixel_width - pwidth + (pwidth * (p/length)); points[pi].y = effective_height - ((*x)->value * (effective_height - 1.)); } /* draw the line */ redraw_end_xfade_to (ar, width, points, effective_height, handle_right, pwidth); /* ensure trim handle stays on top */ if (frame_handle_end) { frame_handle_end->raise_to_top(); } } samplepos_t AudioRegionView::get_fade_in_shape_width () { return audio_region()->fade_in()->back()->when.samples(); } samplepos_t AudioRegionView::get_fade_out_shape_width () { return audio_region()->fade_out()->back()->when.samples(); } void AudioRegionView::redraw_start_xfade () { std::shared_ptr ar (audio_region()); if (!ar->fade_in() || ar->fade_in()->empty()) { return; } show_start_xfade(); reset_fade_in_shape_width (ar, ar->fade_in()->back()->when.samples()); } void AudioRegionView::redraw_start_xfade_to (std::shared_ptr ar, samplecnt_t /*width*/, Points& points, double effective_height, double rect_width) { if (points.size() < 2) { return; } if (!start_xfade_curve) { start_xfade_curve = new ArdourCanvas::XFadeCurve (group, ArdourCanvas::XFadeCurve::Start); CANVAS_DEBUG_NAME (start_xfade_curve, string_compose ("xfade start out line for %1", region()->name())); start_xfade_curve->set_fill_color (UIConfiguration::instance().color_mod ("active crossfade", "crossfade alpha")); start_xfade_curve->set_outline_color (UIConfiguration::instance().color ("crossfade line")); start_xfade_curve->set_ignore_events (true); } if (!start_xfade_rect) { start_xfade_rect = new ArdourCanvas::Rectangle (group); CANVAS_DEBUG_NAME (start_xfade_rect, string_compose ("xfade start rect for %1", region()->name())); start_xfade_rect->set_outline_color (UIConfiguration::instance().color ("crossfade line")); start_xfade_rect->set_fill (false); start_xfade_rect->set_outline (false); start_xfade_rect->Event.connect (sigc::bind (sigc::mem_fun (PublicEditor::instance(), &PublicEditor::canvas_start_xfade_event), start_xfade_rect, this)); start_xfade_rect->set_data ("regionview", this); } start_xfade_rect->set (ArdourCanvas::Rect (0.0, 0.0, rect_width, effective_height)); /* fade out line */ std::shared_ptr inverse = ar->inverse_fade_in (); Points ipoints; Points::size_type npoints; if (!inverse) { /* there is no explicit inverse fade in curve, so take the * regular fade in curve given to use as "points" (already a * set of coordinates), and convert to the inverse shape. */ npoints = points.size(); ipoints.assign (npoints, Duple()); for (Points::size_type i = 0, pci = 0; i < npoints; ++i, ++pci) { ArdourCanvas::Duple &p (ipoints[pci]); /* leave x-axis alone but invert with respect to y-axis */ p.y = effective_height - points[pci].y; } } else { /* there is an explicit inverse fade in curve. Grab the points and convert them into coordinates for the inverse fade in line. */ npoints = inverse->size(); ipoints.assign (npoints, Duple()); Evoral::ControlList::const_iterator x; Points::size_type pi; const double length = inverse->length().samples(); for (x = inverse->begin(), pi = 0; x != inverse->end(); ++x, ++pi) { ArdourCanvas::Duple& p (ipoints[pi]); double pos = (*x)->when.samples(); p.x = (rect_width * (pos/length)); p.y = effective_height - ((*x)->value * (effective_height)); } } start_xfade_curve->set_inout (points, ipoints); show_start_xfade(); } void AudioRegionView::redraw_end_xfade () { std::shared_ptr ar (audio_region()); if (!ar->fade_out() || ar->fade_out()->empty()) { return; } show_end_xfade(); reset_fade_out_shape_width (ar, ar->fade_out()->back()->when.samples()); } void AudioRegionView::redraw_end_xfade_to (std::shared_ptr ar, samplecnt_t width, Points& points, double effective_height, double rect_edge, double rect_width) { if (points.size() < 2) { return; } if (!end_xfade_curve) { end_xfade_curve = new ArdourCanvas::XFadeCurve (group, ArdourCanvas::XFadeCurve::End); CANVAS_DEBUG_NAME (end_xfade_curve, string_compose ("xfade end out line for %1", region()->name())); end_xfade_curve->set_fill_color (UIConfiguration::instance().color_mod ("active crossfade", "crossfade alpha")); end_xfade_curve->set_outline_color (UIConfiguration::instance().color ("crossfade line")); end_xfade_curve->set_ignore_events (true); } if (!end_xfade_rect) { end_xfade_rect = new ArdourCanvas::Rectangle (group); CANVAS_DEBUG_NAME (end_xfade_rect, string_compose ("xfade end rect for %1", region()->name())); end_xfade_rect->set_outline_color (UIConfiguration::instance().color ("crossfade line")); end_xfade_rect->set_fill (false); end_xfade_rect->set_outline (false); end_xfade_rect->Event.connect (sigc::bind (sigc::mem_fun (PublicEditor::instance(), &PublicEditor::canvas_end_xfade_event), end_xfade_rect, this)); end_xfade_rect->set_data ("regionview", this); } end_xfade_rect->set (ArdourCanvas::Rect (rect_edge, 0.0, rect_edge + rect_width, effective_height)); /* fade in line */ std::shared_ptr inverse = ar->inverse_fade_out (); Points ipoints; Points::size_type npoints; if (!inverse) { /* there is no explicit inverse fade out curve, so take the * regular fade out curve given to use as "points" (already a * set of coordinates), and convert to the inverse shape. */ npoints = points.size(); ipoints.assign (npoints, Duple()); Points::size_type pci; for (pci = 0; pci < npoints; ++pci) { ArdourCanvas::Duple &p (ipoints[pci]); p.y = effective_height - points[pci].y; } } else { /* there is an explicit inverse fade out curve. Grab the points and convert them into coordinates for the inverse fade out line. */ npoints = inverse->size(); ipoints.assign (npoints, Duple()); const double rend = trackview.editor().sample_to_pixel (_region->length_samples() - width); Evoral::ControlList::const_iterator x; Points::size_type pi; const double length = inverse->length().samples(); for (x = inverse->begin(), pi = 0; x != inverse->end(); ++x, ++pi) { ArdourCanvas::Duple& p (ipoints[pi]); const double pos = (*x)->when.samples(); p.x = (rect_width * (pos/length)) + rend; p.y = effective_height - ((*x)->value * (effective_height)); } } end_xfade_curve->set_inout (ipoints, points); show_end_xfade(); } void AudioRegionView::hide_xfades () { hide_start_xfade (); hide_end_xfade (); } void AudioRegionView::hide_start_xfade () { if (start_xfade_curve) { start_xfade_curve->hide(); } if (start_xfade_rect) { start_xfade_rect->hide (); } _start_xfade_visible = false; } void AudioRegionView::hide_end_xfade () { if (end_xfade_curve) { end_xfade_curve->hide(); } if (end_xfade_rect) { end_xfade_rect->hide (); } _end_xfade_visible = false; } void AudioRegionView::show_start_xfade () { if (start_xfade_curve) { start_xfade_curve->show(); } if (start_xfade_rect) { start_xfade_rect->show (); } _start_xfade_visible = true; } void AudioRegionView::show_end_xfade () { if (end_xfade_curve) { end_xfade_curve->show(); } if (end_xfade_rect) { end_xfade_rect->show (); } _end_xfade_visible = true; } void AudioRegionView::set_samples_per_pixel (gdouble fpp) { RegionView::set_samples_per_pixel (fpp); if (UIConfiguration::instance().get_show_waveforms ()) { for (uint32_t n = 0; n < waves.size(); ++n) { waves[n]->set_samples_per_pixel (fpp); } } if (_fx_line) { _fx_line->reset (); } reset_fade_shapes (); } void AudioRegionView::set_amplitude_above_axis (gdouble a) { for (uint32_t n=0; n < waves.size(); ++n) { waves[n]->set_amplitude_above_axis (a); } } void AudioRegionView::set_colors () { RegionView::set_colors(); set_fx_line_colors (); set_waveform_colors (); if (start_xfade_curve) { start_xfade_curve->set_fill_color (UIConfiguration::instance().color_mod ("active crossfade", "crossfade alpha")); start_xfade_curve->set_outline_color (UIConfiguration::instance().color ("crossfade line")); } if (end_xfade_curve) { end_xfade_curve->set_fill_color (UIConfiguration::instance().color_mod ("active crossfade", "crossfade alpha")); end_xfade_curve->set_outline_color (UIConfiguration::instance().color ("crossfade line")); } if (start_xfade_rect) { start_xfade_rect->set_outline_color (UIConfiguration::instance().color ("crossfade line")); } if (end_xfade_rect) { end_xfade_rect->set_outline_color (UIConfiguration::instance().color ("crossfade line")); } } void AudioRegionView::setup_waveform_visibility () { if (UIConfiguration::instance().get_show_waveforms ()) { for (uint32_t n = 0; n < waves.size(); ++n) { /* make sure the zoom level is correct, since we don't update this when waveforms are hidden. */ // CAIROCANVAS // waves[n]->set_samples_per_pixel (_samples_per_pixel); waves[n]->show(); } } else { for (uint32_t n = 0; n < waves.size(); ++n) { waves[n]->hide(); } } } void AudioRegionView::temporarily_hide_envelope () { if (_fx_line) { _fx_line->hide (); } } void AudioRegionView::unhide_envelope () { update_envelope_visibility (); } void AudioRegionView::set_region_gain_line () { if (_ignore_line_change) { return; } const string line_name = _region->name() + ":gain"; _fx_line.reset (new AudioRegionGainLine (line_name, *this, *group, audio_region()->envelope())); _fx_line->set_height ((uint32_t) rint (height() - NAME_HIGHLIGHT_SIZE) - 2); _fx_line->reset (); _region_fx_connection.disconnect (); bool changed = _rfx_id != PBD::ID (0) || _rdx_param != UINT32_MAX; _rfx_id = PBD::ID (0); _rdx_param = UINT32_MAX; envelope_active_changed (); if (changed) { region_line_changed (); /* EMIT SIGNAL */ } } void AudioRegionView::set_region_fx_line (std::shared_ptr ac, std::shared_ptr rfx, uint32_t param_id) { const string line_name = _region->name () + ":" + rfx->describe_parameter (Evoral::Parameter (PluginAutomation, 0, param_id)); _fx_line.reset (new RegionFxLine (line_name, *this, *group, ac)); _fx_line->set_height ((uint32_t) rint (height() - NAME_HIGHLIGHT_SIZE) - 2); _fx_line->reset (); rfx->DropReferences.connect (_region_fx_connection, invalidator (*this), boost::bind (&AudioRegionView::set_region_gain_line, this), gui_context ()); bool changed = _rfx_id != rfx->id () || _rdx_param != param_id; _rfx_id = rfx->id (); _rdx_param = param_id; envelope_active_changed (); if (changed) { region_line_changed (); /* EMIT SIGNAL */ } } bool AudioRegionView::set_region_fx_line (uint32_t plugin_id, uint32_t param_id) { if (_ignore_line_change) { return false; } std::shared_ptr rfx = _region->nth_plugin (plugin_id); if (rfx) { std::shared_ptr c = rfx->control (Evoral::Parameter (PluginAutomation, 0, param_id)); std::shared_ptr ac = std::dynamic_pointer_cast (c); if (ac) { set_region_fx_line (ac, rfx, param_id); return true; } } return false; } bool AudioRegionView::set_region_fx_line (std::weak_ptr wac) { if (_ignore_line_change) { return false; } std::shared_ptr ac = std::dynamic_pointer_cast (wac.lock ()); if (!ac) { return false; } bool found = false; _region->foreach_plugin ([this, ac, &found](std::weak_ptr wfx) { std::shared_ptr rfx (wfx.lock ()); if (!rfx || found) { return; } std::shared_ptr plugin = rfx->plugin (); for (size_t i = 0; i < plugin->parameter_count (); ++i) { if (!plugin->parameter_is_control (i) || !plugin->parameter_is_input (i)) { continue; } const Evoral::Parameter param (PluginAutomation, 0, i); if (ac == std::dynamic_pointer_cast (rfx->control (param))) { set_region_fx_line (ac, rfx, i); found = true; break; } } }); return found; } bool AudioRegionView::get_region_fx_line (PBD::ID& id, uint32_t& param_id) { id = _rfx_id; param_id = _rdx_param; return _rdx_param != UINT32_MAX && _rfx_id != 0; } bool AudioRegionView::paste (Temporal::timepos_t const& pos, const Selection& selection, PasteContext& ctx) { if (!_fx_line || selection.lines.size() != 1) { return false; } std::shared_ptr alist (_fx_line->the_list()); AutomationSelection::const_iterator p = selection.lines.begin (); Temporal::timecnt_t len = (*p)->length(); Temporal::timepos_t tpos (pos); unsigned paste_count = ctx.count; switch (alist->time_domain()) { case Temporal::BeatTime: tpos += trackview.editor().get_paste_offset (pos, paste_count > 0 ? 1 : 0, len); break; case Temporal::AudioTime: tpos += trackview.editor().get_paste_offset (pos, paste_count, len); break; } XMLNode &before = alist->get_state(); Temporal::timepos_t model_pos (_region->position().distance (tpos)); alist->paste (**p, model_pos); timepos_t rlen ((samplepos_t)_region->length().samples()); alist->truncate_end (rlen); trackview.session()->add_command (new MementoCommand(*alist.get(), &before, &alist->get_state())); return true; } void AudioRegionView::update_envelope_visibility () { if (!_fx_line) { return; } if (trackview.editor().current_mouse_mode() == Editing::MouseDraw || trackview.editor().current_mouse_mode() == Editing::MouseContent ) { _fx_line->set_visibility (AutomationLine::VisibleAspects(AutomationLine::ControlPoints|AutomationLine::Line)); _fx_line->canvas_group().raise_to_top (); } else if (UIConfiguration::instance().get_show_region_gain() || trackview.editor().current_mouse_mode() == Editing::MouseRange ) { _fx_line->set_visibility (AutomationLine::VisibleAspects(AutomationLine::Line)); _fx_line->canvas_group().raise_to_top (); } else { _fx_line->set_visibility (AutomationLine::VisibleAspects(0)); } } void AudioRegionView::delete_waves () { for (vector::iterator i = _data_ready_connections.begin(); i != _data_ready_connections.end(); ++i) { delete *i; } _data_ready_connections.clear (); for (vector::iterator w = waves.begin(); w != waves.end(); ++w) { group->remove(*w); delete *w; } waves.clear(); while (!tmp_waves.empty ()) { delete tmp_waves.back (); tmp_waves.pop_back (); } pending_peak_data->show (); } void AudioRegionView::create_waves () { // cerr << "AudioRegionView::create_waves() called on " << this << endl;//DEBUG RouteTimeAxisView& atv (*(dynamic_cast(&trackview))); // ick if (!atv.track()) { return; } ChanCount nchans = atv.track()->n_channels(); // cerr << "creating waves for " << _region->name() << " with wfd = " << wait_for_data // << " and channels = " << nchans.n_audio() << endl; /* in tmp_waves, set up null pointers for each channel so the vector is allocated */ for (uint32_t n = 0; n < nchans.n_audio(); ++n) { tmp_waves.push_back (0); } for (vector::iterator i = _data_ready_connections.begin(); i != _data_ready_connections.end(); ++i) { delete *i; } _data_ready_connections.clear (); for (uint32_t i = 0; i < nchans.n_audio(); ++i) { _data_ready_connections.push_back (0); } for (uint32_t n = 0; n < nchans.n_audio(); ++n) { if (n >= audio_region()->n_channels()) { break; } // cerr << "\tchannel " << n << endl; if (wait_for_data) { if (audio_region()->audio_source(n)->peaks_ready (boost::bind (&AudioRegionView::peaks_ready_handler, this, n), &_data_ready_connections[n], gui_context())) { // cerr << "\tData is ready for channel " << n << "\n"; create_one_wave (n, true); } else { // cerr << "\tdata is not ready for channel " << n << "\n"; // we'll get a PeaksReady signal from the source in the future // and will call create_one_wave(n) then. pending_peak_data->show (); } } else { // cerr << "\tdon't delay, display channel " << n << " today!\n"; create_one_wave (n, true); } } } void AudioRegionView::create_one_wave (uint32_t which, bool /*direct*/) { //cerr << "AudioRegionView::create_one_wave() called which: " << which << " this: " << this << endl;//DEBUG RouteTimeAxisView& atv (*(dynamic_cast(&trackview))); // ick if (!trackview.session() || trackview.session()->deletion_in_progress () || !atv.track()) { /* peaks_ready_handler() may be called from peak_thread_work() while * session deletion is in progress. * Since session-unload happens in the GUI thread, we need to test * in this context. */ return; } uint32_t nchans = atv.track()->n_channels().n_audio(); uint32_t n; uint32_t nwaves = std::min (nchans, audio_region()->n_channels()); gdouble ht; /* compare to set_height(), use _height as set by streamview (child_height), * not trackview.current_height() to take stacked layering into acconnt */ if (!UIConfiguration::instance().get_show_name_highlight() || (_height < NAME_HIGHLIGHT_THRESH)) { ht = _height / (double) nchans; } else { ht = (_height - NAME_HIGHLIGHT_SIZE) / (double) nchans; } /* first waveview starts at 1.0, not 0.0 since that will overlap the frame */ gdouble yoff = which * ht; ArdourWaveView::WaveView *wave = new ArdourWaveView::WaveView (group, audio_region ()); CANVAS_DEBUG_NAME (wave, string_compose ("wave view for chn %1 of %2", which, get_item_name())); wave->set_channel (which); wave->set_y_position (yoff); wave->set_height (ht); wave->set_samples_per_pixel (samples_per_pixel); wave->set_show_zero_line (true); wave->set_clip_level (UIConfiguration::instance().get_waveform_clip_level ()); wave->set_start_shift (1.0); wave->Event.connect (sigc::bind (sigc::mem_fun (PublicEditor::instance(), &PublicEditor::canvas_wave_view_event), wave, this)); switch (UIConfiguration::instance().get_waveform_shape()) { case Rectified: wave->set_shape (ArdourWaveView::WaveView::Rectified); break; default: wave->set_shape (ArdourWaveView::WaveView::Normal); } wave->set_logscaled (UIConfiguration::instance().get_waveform_scale() == Logarithmic); vector v; v.push_back (wave); set_some_waveform_colors (v); if (!UIConfiguration::instance().get_show_waveforms ()) { wave->hide(); } /* note: calling this function is serialized by the lock held in the peak building thread that signals that peaks are ready for use *or* by the fact that it is called one by one from the GUI thread. */ if (which < nchans) { tmp_waves[which] = wave; } else { /* n-channel track, >n-channel source */ } /* see if we're all ready */ for (n = 0; n < nchans; ++n) { if (tmp_waves[n] == 0) { break; } } if (n == nwaves) { /* all waves are ready */ tmp_waves.resize(nwaves); waves.swap(tmp_waves); while (!tmp_waves.empty ()) { delete tmp_waves.back (); tmp_waves.pop_back (); } /* indicate peak-completed */ pending_peak_data->hide (); /* Restore stacked coverage */ LayerDisplay layer_display; if (trackview.get_gui_property ("layer-display", layer_display)) { update_coverage_frame (layer_display); } } /* channel wave created, don't hook into peaks ready anymore */ delete _data_ready_connections[which]; _data_ready_connections[which] = 0; maybe_raise_cue_markers (); } void AudioRegionView::peaks_ready_handler (uint32_t which) { Gtkmm2ext::UI::instance()->call_slot (invalidator (*this), boost::bind (&AudioRegionView::create_one_wave, this, which, false)); // cerr << "AudioRegionView::peaks_ready_handler() called on " << which << " this: " << this << endl; } void AudioRegionView::add_gain_point_event (ArdourCanvas::Item *item, GdkEvent *ev, bool with_guard_points) { if (!_fx_line) { return; } uint32_t before_p, after_p; double mx = ev->button.x; double my = ev->button.y; item->canvas_to_item (mx, my); samplecnt_t const sample_within_region = (samplecnt_t) floor (mx * samples_per_pixel); double y = my; if (_fx_line->control_points_adjacent (sample_within_region, before_p, after_p)) { /* y is in item frame */ double const bx = _fx_line->nth (before_p)->get_x(); double const ax = _fx_line->nth (after_p)->get_x(); double const click_ratio = (ax - mx) / (ax - bx); y = ((_fx_line->nth (before_p)->get_y() * click_ratio) + (_fx_line->nth (after_p)->get_y() * (1 - click_ratio))); } /* don't create points that can't be seen */ update_envelope_visibility (); samplepos_t rpos = region ()->position_sample(); timepos_t snap_pos = timepos_t (trackview.editor().pixel_to_sample (mx) + rpos); trackview.editor ().snap_to_with_modifier (snap_pos, ev); samplepos_t fx = snap_pos.samples() - rpos; if (fx > _region->length_samples()) { return; } /* compute vertical fractional position */ y = 1.0 - (y / (_fx_line->height())); /* map using gain line */ _fx_line->view_to_model_coord_y (y); /* XXX STATEFUL: can't convert to stateful diff until we can represent automation data with it. */ XMLNode &before = _fx_line->the_list()->get_state(); if (_fx_line->the_list()->editor_add (timepos_t (fx), y, with_guard_points)) { XMLNode &after = _fx_line->the_list()->get_state(); std::list results; trackview.editor().begin_reversible_command (_("add gain control point")); _fx_line->enable_automation (); trackview.session()->add_command (new MementoCommand(*_fx_line->the_list(), &before, &after)); _fx_line->get_selectables (region ()->position () + timecnt_t (fx), region ()->position () + timecnt_t (fx), 0.0, 1.0, results); trackview.editor ().get_selection ().set (results); trackview.editor ().commit_reversible_command (); trackview.session ()->set_dirty (); } else { delete &before; } } GhostRegion* AudioRegionView::add_ghost (TimeAxisView& tv) { RouteTimeAxisView* rtv = dynamic_cast(&trackview); if (!rtv) { return 0; } double unit_position = _region->position_sample () / samples_per_pixel; AudioGhostRegion* ghost = new AudioGhostRegion (*this, tv, trackview, unit_position); uint32_t nchans; nchans = rtv->track()->n_channels().n_audio(); for (uint32_t n = 0; n < nchans; ++n) { if (n >= audio_region()->n_channels()) { break; } ArdourWaveView::WaveView *wave = new ArdourWaveView::WaveView (ghost->group, audio_region()); CANVAS_DEBUG_NAME (wave, string_compose ("ghost wave for %1", get_item_name())); wave->set_channel (n); wave->set_samples_per_pixel (samples_per_pixel); wave->set_amplitude_above_axis (_amplitude_above_axis); ghost->waves.push_back(wave); } ghost->set_height (); ghost->set_duration (_region->length_samples() / samples_per_pixel); ghost->set_colors(); ghosts.push_back (ghost); return ghost; } void AudioRegionView::entered () { trackview.editor().set_current_trimmable (_region); trackview.editor().set_current_movable (_region); update_envelope_visibility(); if ((trackview.editor().current_mouse_mode() == Editing::MouseObject)) { if (start_xfade_rect) { start_xfade_rect->set_outline (true); } if (end_xfade_rect) { end_xfade_rect->set_outline (true); } if (fade_in_handle) { fade_in_handle->show (); fade_in_handle->raise_to_top (); } if (fade_out_handle) { fade_out_handle->show (); fade_out_handle->raise_to_top (); } if (fade_in_trim_handle) { std::shared_ptr ar (audio_region()); if (!ar->locked() && (ar->fade_in()->back()->when > 64 || (ar->can_trim() & Trimmable::FrontTrimEarlier))) { fade_in_trim_handle->show (); fade_in_trim_handle->raise_to_top (); } else { fade_in_trim_handle->hide (); } } if (fade_out_trim_handle) { std::shared_ptr ar (audio_region()); if (!ar->locked() && (ar->fade_out()->back()->when > 64 || (ar->can_trim() & Trimmable::EndTrimLater))) { fade_out_trim_handle->show (); fade_out_trim_handle->raise_to_top (); } else { fade_out_trim_handle->hide (); } } } else { //this happens when we switch tools; if we switch away from Grab mode, hide all the fade handles if (fade_in_handle) { fade_in_handle->hide(); } if (fade_out_handle) { fade_out_handle->hide(); } if (fade_in_trim_handle) { fade_in_trim_handle->hide(); } if (fade_out_trim_handle) { fade_out_trim_handle->hide(); } if (start_xfade_rect) { start_xfade_rect->set_outline (false); } if (end_xfade_rect) { end_xfade_rect->set_outline (false); } } } void AudioRegionView::exited () { trackview.editor().set_current_trimmable (std::shared_ptr()); trackview.editor().set_current_movable (std::shared_ptr()); if (fade_in_handle) { fade_in_handle->hide(); } if (fade_out_handle) { fade_out_handle->hide(); } if (fade_in_trim_handle) { fade_in_trim_handle->hide(); } if (fade_out_trim_handle) { fade_out_trim_handle->hide(); } if (start_xfade_rect) { start_xfade_rect->set_outline (false); } if (end_xfade_rect) { end_xfade_rect->set_outline (false); } } void AudioRegionView::envelope_active_changed () { set_fx_line_colors (); update_envelope_visibility (); } void AudioRegionView::color_handler () { //case cMutedWaveForm: //case cWaveForm: //case cWaveFormClip: //case cZeroLine: set_colors (); //case cGainLineInactive: //case cGainLine: envelope_active_changed(); } void AudioRegionView::set_fx_line_colors () { if (!_fx_line) { return; } if (_rdx_param != UINT32_MAX && _rfx_id != 0) { _fx_line->set_line_color ("region automation line"); } else if (audio_region()->envelope_active()) { _fx_line->set_line_color ("gain line"); } else { _fx_line->set_line_color ("gain line inactive", "gain line inactive"); } } void AudioRegionView::set_waveform_colors () { set_some_waveform_colors (waves); } void AudioRegionView::set_some_waveform_colors (vector& waves_to_color) { Gtkmm2ext::Color fill = fill_color; Gtkmm2ext::Color outline = fill; Gtkmm2ext::Color clip = UIConfiguration::instance().color ("clipped waveform"); Gtkmm2ext::Color zero = UIConfiguration::instance().color ("zero line"); /* use track/region color to fill wform */ fill = fill_color; fill = UINT_INTERPOLATE (fill, UIConfiguration::instance().color ("waveform fill"), 0.5); /* set outline */ outline = UIConfiguration::instance().color ("waveform outline"); if (_dragging) { outline = UINT_RGBA_CHANGE_A(outline, 0xC0); fill = UINT_RGBA_CHANGE_A(fill, 0xC0); } else if (_selected) { outline = UINT_RGBA_CHANGE_A(UIConfiguration::instance().color ("selected waveform outline"), 0xC0); fill = UINT_RGBA_CHANGE_A(UIConfiguration::instance().color ("selected waveform fill"), 0xC0); } else if (_region->muted()) { outline = UINT_RGBA_CHANGE_A(outline, 0x40); fill = UINT_RGBA_CHANGE_A (UINT_INTERPOLATE(fill_color, UIConfiguration::instance().color ("covered region"), 0.7), 0x40); } else if (!_region->opaque() && trackview.layer_display () != Stacked) { outline = UINT_RGBA_CHANGE_A(outline, 0x80); fill = UINT_RGBA_CHANGE_A(UIConfiguration::instance().color ("waveform fill"), 0x80); } /* recorded region, override to red */ if (_recregion) { outline = UIConfiguration::instance().color ("recording waveform outline"); fill = UIConfiguration::instance().color ("recording waveform fill"); } for (vector::iterator w = waves_to_color.begin(); w != waves_to_color.end(); ++w) { (*w)->set_fill_color (fill); (*w)->set_outline_color (outline); (*w)->set_clip_color (clip); (*w)->set_zero_color (zero); } } void AudioRegionView::set_frame_color () { if (!frame) { return; } RegionView::set_frame_color (); set_waveform_colors (); } void AudioRegionView::set_fade_visibility (bool yn) { if (yn) { if (start_xfade_curve) { start_xfade_curve->show (); } if (end_xfade_curve) { end_xfade_curve->show (); } if (start_xfade_rect) { start_xfade_rect->show (); } if (end_xfade_rect) { end_xfade_rect->show (); } } else { if (start_xfade_curve) { start_xfade_curve->hide(); } if (end_xfade_curve) { end_xfade_curve->hide(); } if (fade_in_handle) { fade_in_handle->hide(); } if (fade_out_handle) { fade_out_handle->hide(); } if (fade_in_trim_handle) { fade_in_trim_handle->hide(); } if (fade_out_trim_handle) { fade_out_trim_handle->hide(); } if (start_xfade_rect) { start_xfade_rect->hide (); } if (end_xfade_rect) { end_xfade_rect->hide (); } if (start_xfade_rect) { start_xfade_rect->set_outline (false); } if (end_xfade_rect) { end_xfade_rect->set_outline (false); } } } void AudioRegionView::update_coverage_frame (LayerDisplay d) { RegionView::update_coverage_frame (d); if (d == Stacked) { if (fade_in_handle) { fade_in_handle->raise_to_top (); } if (fade_out_handle) { fade_out_handle->raise_to_top (); } if (fade_in_trim_handle) { fade_in_trim_handle->raise_to_top (); } if (fade_out_trim_handle) { fade_out_trim_handle->raise_to_top (); } } } void AudioRegionView::show_region_editor () { if (editor == 0) { editor = new AudioRegionEditor (trackview.session(), this); } editor->present (); editor->show_all(); } void AudioRegionView::transients_changed () { AnalysisFeatureList analysis_features; _region->transients (analysis_features); samplepos_t position = _region->position_sample(); samplepos_t first = _region->first_sample(); samplepos_t last = _region->last_sample(); double y1; if (_height >= NAME_HIGHLIGHT_THRESH) { y1 = _height - TimeAxisViewItem::NAME_HIGHLIGHT_SIZE - 1; } else { y1 = _height - 1; } while (feature_lines.size() < analysis_features.size()) { ArdourCanvas::Line* canvas_item = new ArdourCanvas::Line(group); CANVAS_DEBUG_NAME (canvas_item, string_compose ("transient group for %1", region()->name())); canvas_item->set_outline_color (UIConfiguration::instance().color ("zero line")); // also in Editor::leave_handler() canvas_item->set (ArdourCanvas::Duple (-1.0, 2.0), ArdourCanvas::Duple (1.0, y1)); canvas_item->raise_to_top (); canvas_item->show (); canvas_item->set_data ("regionview", this); canvas_item->Event.connect (sigc::bind (sigc::mem_fun (PublicEditor::instance(), &PublicEditor::canvas_feature_line_event), canvas_item, this)); feature_lines.push_back (make_pair(0, canvas_item)); } while (feature_lines.size() > analysis_features.size()) { ArdourCanvas::Line* line = feature_lines.back().second; feature_lines.pop_back (); delete line; } AnalysisFeatureList::const_iterator i; list >::iterator l; for (i = analysis_features.begin(), l = feature_lines.begin(); i != analysis_features.end() && l != feature_lines.end(); ++i, ++l) { float *pos = new float; *pos = trackview.editor().sample_to_pixel (*i - position); (*l).second->set ( ArdourCanvas::Duple (*pos, 2.0), ArdourCanvas::Duple (*pos, y1) ); (*l).second->set_data ("position", pos); // is this *modified* (drag?), if not use *i (*l).first = *i; if (l->first < first || l->first >= last) { l->second->hide(); } else { l->second->show(); } } } void AudioRegionView::update_transient(float /*old_pos*/, float new_pos) { /* Find sample at old pos, calculate new sample then update region transients*/ list >::iterator l; for (l = feature_lines.begin(); l != feature_lines.end(); ++l) { /* Line has been updated in drag so we compare to new_pos */ float* pos = (float*) (*l).second->get_data ("position"); if (rint(new_pos) == rint(*pos)) { samplepos_t position = _region->position_sample(); samplepos_t old_sample = (*l).first; samplepos_t new_sample = trackview.editor().pixel_to_sample (new_pos) + position; _region->update_transient (old_sample, new_sample); break; } } } void AudioRegionView::remove_transient (float pos) { /* this is called from Editor::remove_transient () with pos == get_data ("position") * which is the item's x-coordinate inside the ARV. * * Find sample at old pos, calculate new sample then update region transients */ list >::iterator l; for (l = feature_lines.begin(); l != feature_lines.end(); ++l) { float *line_pos = (float*) (*l).second->get_data ("position"); if (rint(pos) == rint(*line_pos)) { _region->remove_transient ((*l).first); break; } } } void AudioRegionView::thaw_after_trim () { RegionView::thaw_after_trim (); unhide_envelope (); drag_end (); } void AudioRegionView::show_xfades () { show_start_xfade (); show_end_xfade (); } void AudioRegionView::drag_start () { RegionView::drag_start (); //we used to hide xfades here. I don't see the point with the new model, but we can re-implement if needed } void AudioRegionView::drag_end () { TimeAxisViewItem::drag_end (); //see comment for drag_start if (fade_in_handle && fade_in_handle->visible()) { // length of region or fade changed, re-check // if fade_in_trim_handle or fade_out_trim_handle should // be visible. -- If the fade_in_handle is visible // we have focus and are not in internal edit mode. entered(); } } void AudioRegionView::parameter_changed (string const & p) { RegionView::parameter_changed (p); if (p == "show-waveforms") { setup_waveform_visibility (); } } MergeableLine* AudioRegionView::make_merger () { return new MergeableLine (_fx_line, std::shared_ptr(), [this](timepos_t const& t) { return timepos_t (_region->position().distance (t)); }, nullptr, nullptr); }