session-ios/Signal/src/Models/TSMessageAdapaters/TSVideoAttachmentAdapter.m

447 lines
15 KiB
Mathematica
Raw Normal View History

//
// Copyright (c) 2017 Open Whisper Systems. All rights reserved.
//
#import "TSVideoAttachmentAdapter.h"
#import "AttachmentUploadView.h"
#import "JSQMediaItem+OWS.h"
#import "MIMETypeUtil.h"
#import "Signal-Swift.h"
#import "TSAttachmentStream.h"
#import "TSMessagesManager.h"
#import "TSStorageManager+keyingMaterial.h"
#import "UIColor+JSQMessages.h"
#import "UIColor+OWS.h"
#import "UIFont+OWS.h"
#import "UIView+OWS.h"
#import "ViewControllerUtils.h"
#import <JSQMessagesViewController/JSQMessagesMediaViewBubbleImageMasker.h>
#import <MobileCoreServices/MobileCoreServices.h>
#import <SignalServiceKit/MIMETypeUtil.h>
2017-03-29 20:49:26 +02:00
NS_ASSUME_NONNULL_BEGIN
@interface TSVideoAttachmentAdapter () <AudioProgressViewDelegate>
@property (nonatomic) UIImage *image;
@property (nonatomic, nullable) UIView *cachedMediaView;
@property (nonatomic) TSAttachmentStream *attachment;
2017-03-29 20:49:26 +02:00
@property (nonatomic, nullable) UIButton *audioPlayPauseButton;
@property (nonatomic, nullable) UILabel *audioBottomLabel;
@property (nonatomic, nullable) AudioProgressView *audioProgressView;
2015-01-25 00:48:40 +01:00
@property (nonatomic) BOOL incoming;
2017-03-29 20:49:26 +02:00
@property (nonatomic, nullable) AttachmentUploadView *attachmentUploadView;
@property (nonatomic) BOOL isAudioPlaying;
@property (nonatomic) CGFloat audioProgressSeconds;
@property (nonatomic) CGFloat audioDurationSeconds;
@property (nonatomic) BOOL isPaused;
2017-04-21 20:19:30 +02:00
// See comments on OWSMessageMediaAdapter.
@property (nonatomic, nullable, weak) id lastPresentingCell;
@end
2017-03-29 20:49:26 +02:00
#pragma mark -
@implementation TSVideoAttachmentAdapter
- (instancetype)initWithAttachment:(TSAttachmentStream *)attachment incoming:(BOOL)incoming {
self = [super initWithFileURL:[attachment mediaURL] isReadyToPlay:YES];
if (self) {
_image = attachment.image;
_cachedMediaView = nil;
_attachmentId = attachment.uniqueId;
_contentType = attachment.contentType;
2015-01-25 00:48:40 +01:00
_attachment = attachment;
_incoming = incoming;
}
return self;
}
- (void)clearAllViews
{
[self.cachedMediaView removeFromSuperview];
self.cachedMediaView = nil;
self.attachmentUploadView = nil;
self.audioProgressView = nil;
}
- (void)clearCachedMediaViews
{
[super clearCachedMediaViews];
[self clearAllViews];
}
- (void)setAppliesMediaViewMaskAsOutgoing:(BOOL)appliesMediaViewMaskAsOutgoing
{
[super setAppliesMediaViewMaskAsOutgoing:appliesMediaViewMaskAsOutgoing];
[self clearAllViews];
}
- (BOOL)isAudio {
return [MIMETypeUtil isSupportedAudioMIMEType:_contentType];
}
- (BOOL)isVideo {
return [MIMETypeUtil isSupportedVideoMIMEType:_contentType];
}
- (void)setAudioProgress:(CGFloat)progress duration:(CGFloat)duration
{
OWSAssert([NSThread isMainThread]);
self.audioProgressSeconds = progress;
if (duration > 0) {
self.audioDurationSeconds = duration;
}
[self updateAudioProgressView];
[self updateAudioBottomLabel];
}
- (void)updateAudioBottomLabel
{
if (self.isAudioPlaying && self.audioProgressSeconds > 0 && self.audioDurationSeconds > 0) {
2017-05-03 16:50:26 +02:00
self.audioBottomLabel.text =
[NSString stringWithFormat:@"%@ / %@",
[ViewControllerUtils formatDurationSeconds:(long)round(self.audioProgressSeconds)],
[ViewControllerUtils formatDurationSeconds:(long)round(self.audioDurationSeconds)]];
} else {
self.audioBottomLabel.text = [NSString
stringWithFormat:@"%@", [ViewControllerUtils formatDurationSeconds:(long)round(self.audioDurationSeconds)]];
}
}
- (void)setAudioIcon:(UIImage *)icon iconColor:(UIColor *)iconColor
{
icon = [icon imageWithRenderingMode:UIImageRenderingModeAlwaysTemplate];
[_audioPlayPauseButton setImage:icon forState:UIControlStateNormal];
[_audioPlayPauseButton setImage:icon forState:UIControlStateDisabled];
_audioPlayPauseButton.imageView.tintColor = iconColor;
}
2015-01-25 00:48:40 +01:00
- (void)setAudioIconToPlay {
[self setAudioIcon:[UIImage imageNamed:@"audio_play_black_40"]
iconColor:[self audioColorWithOpacity:self.incoming ? 0.2f : 0.1f]];
2015-01-25 00:48:40 +01:00
}
- (void)setAudioIconToPause {
[self setAudioIcon:[UIImage imageNamed:@"audio_pause_black_40"]
iconColor:[self audioColorWithOpacity:self.incoming ? 0.2f : 0.1f]];
}
- (void)setIsAudioPlaying:(BOOL)isAudioPlaying
{
_isAudioPlaying = isAudioPlaying;
[self updateAudioProgressView];
}
- (void)updateAudioProgressView
{
[self.audioProgressView
setProgress:(self.audioDurationSeconds > 0 ? self.audioProgressSeconds / self.audioDurationSeconds : 0.f)];
self.audioProgressView.horizontalBarColor = [self audioColorWithOpacity:0.75f];
self.audioProgressView.progressColor
= (self.isAudioPlaying ? [self audioColorWithOpacity:self.incoming ? 0.2f : 0.1f]
: [self audioColorWithOpacity:0.4f]);
}
#pragma mark - JSQMessageMediaData protocol
- (CGFloat)audioBubbleHeight
{
return 45.f;
}
- (CGFloat)iconSize
{
return 40.f;
}
- (CGFloat)vMargin
{
return 5.f;
}
- (UIColor *)audioTextColor
{
return (self.incoming ? [UIColor colorWithWhite:0.2f alpha:1.f] : [UIColor whiteColor]);
}
- (UIColor *)audioColorWithOpacity:(CGFloat)alpha
{
return [self.audioTextColor blendWithColor:self.bubbleBackgroundColor alpha:alpha];
}
- (UIColor *)bubbleBackgroundColor
{
return self.incoming ? [UIColor jsq_messageBubbleLightGrayColor] : [UIColor ows_materialBlueColor];
}
- (UIView *)mediaView {
if ([self isVideo]) {
if (self.cachedMediaView == nil) {
self.cachedMediaView = [self createVideoMediaView];
}
} else if ([self isAudio]) {
if (self.cachedMediaView == nil) {
self.cachedMediaView = [self createAudioMediaView];
}
if (self.isAudioPlaying) {
[self setAudioIconToPause];
} else {
[self setAudioIconToPlay];
}
} else {
// Unknown media type.
OWSAssert(0);
}
return self.cachedMediaView;
}
- (UIView *)createVideoMediaView
{
OWSAssert([self isVideo]);
CGSize size = [self mediaViewDisplaySize];
UIImageView *imageView = [[UIImageView alloc] initWithImage:self.image];
imageView.contentMode = UIViewContentModeScaleAspectFill;
imageView.frame = CGRectMake(0.0f, 0.0f, size.width, size.height);
imageView.clipsToBounds = YES;
[JSQMessagesMediaViewBubbleImageMasker applyBubbleImageMaskToMediaView:imageView
isOutgoing:self.appliesMediaViewMaskAsOutgoing];
UIImage *img = [UIImage imageNamed:@"play_button"];
UIImageView *videoPlayButton = [[UIImageView alloc] initWithImage:img];
videoPlayButton.frame = CGRectMake((size.width / 2) - 18, (size.height / 2) - 18, 37, 37);
[imageView addSubview:videoPlayButton];
if (!_incoming) {
self.attachmentUploadView = [[AttachmentUploadView alloc] initWithAttachment:self.attachment
superview:imageView
attachmentStateCallback:^(BOOL isAttachmentReady) {
videoPlayButton.hidden = !isAttachmentReady;
}];
}
return imageView;
}
- (BOOL)isVoiceMessage
{
OWSAssert([self isAudio]);
return (self.attachment.isVoiceMessage || self.attachment.filename.length < 1);
}
- (UIView *)createAudioMediaView
{
OWSAssert([self isAudio]);
[self ensureAudioDurationSeconds];
CGSize viewSize = [self mediaViewDisplaySize];
UIColor *textColor = [self audioTextColor];
UIView *mediaView = [[UIView alloc] initWithFrame:CGRectMake(0.f, 0.f, viewSize.width, viewSize.height)];
mediaView.backgroundColor = self.bubbleBackgroundColor;
[JSQMessagesMediaViewBubbleImageMasker applyBubbleImageMaskToMediaView:mediaView isOutgoing:!self.incoming];
const CGFloat kBubbleTailWidth = 6.f;
CGRect contentFrame = CGRectMake(self.incoming ? kBubbleTailWidth : 0.f,
self.vMargin,
viewSize.width - kBubbleTailWidth - 15,
viewSize.height - self.vMargin * 2);
CGRect iconFrame = CGRectMake((CGFloat)round(contentFrame.origin.x + 5.f),
(CGFloat)round(contentFrame.origin.y + (contentFrame.size.height - self.iconSize) * 0.5f),
self.iconSize,
self.iconSize);
_audioPlayPauseButton = [[UIButton alloc] initWithFrame:iconFrame];
_audioPlayPauseButton.enabled = NO;
[mediaView addSubview:_audioPlayPauseButton];
const CGFloat kLabelHSpacing = 3;
const CGFloat kLabelVSpacing = 2;
NSString *topText = [[self.attachment.filename stringByDeletingPathExtension]
stringByTrimmingCharactersInSet:[NSCharacterSet whitespaceCharacterSet]];
if (topText.length < 1) {
topText = [MIMETypeUtil fileExtensionForMIMEType:self.attachment.contentType].uppercaseString;
}
if (topText.length < 1) {
topText = NSLocalizedString(@"GENERIC_ATTACHMENT_LABEL", @"A label for generic attachments.");
}
if (self.isVoiceMessage) {
topText = nil;
}
UILabel *topLabel = [UILabel new];
topLabel.text = topText;
topLabel.textColor = [textColor colorWithAlphaComponent:0.85f];
topLabel.lineBreakMode = NSLineBreakByTruncatingMiddle;
topLabel.font = [UIFont ows_regularFontWithSize:ScaleFromIPhone5To7Plus(11.f, 13.f)];
[topLabel sizeToFit];
[mediaView addSubview:topLabel];
AudioProgressView *audioProgressView = [AudioProgressView new];
self.audioProgressView = audioProgressView;
audioProgressView.delegate = self;
[self updateAudioProgressView];
[mediaView addSubview:audioProgressView];
UILabel *bottomLabel = [UILabel new];
self.audioBottomLabel = bottomLabel;
[self updateAudioBottomLabel];
bottomLabel.textColor = [textColor colorWithAlphaComponent:0.85f];
bottomLabel.lineBreakMode = NSLineBreakByTruncatingMiddle;
bottomLabel.font = [UIFont ows_regularFontWithSize:ScaleFromIPhone5To7Plus(11.f, 13.f)];
[bottomLabel sizeToFit];
[mediaView addSubview:bottomLabel];
const CGFloat topLabelHeight = ceil(topLabel.font.lineHeight);
const CGFloat kAudioProgressViewHeight = 12.f;
const CGFloat bottomLabelHeight = ceil(bottomLabel.font.lineHeight);
CGRect labelsBounds = CGRectZero;
labelsBounds.origin.x = (CGFloat)round(iconFrame.origin.x + iconFrame.size.width + kLabelHSpacing);
labelsBounds.size.width = contentFrame.origin.x + contentFrame.size.width - labelsBounds.origin.x;
labelsBounds.size.height = topLabelHeight + kAudioProgressViewHeight + bottomLabelHeight + kLabelVSpacing * 2;
labelsBounds.origin.y
= (CGFloat)round(contentFrame.origin.y + (contentFrame.size.height - labelsBounds.size.height) * 0.5f);
topLabel.frame = CGRectMake(labelsBounds.origin.x, labelsBounds.origin.y, labelsBounds.size.width, topLabelHeight);
audioProgressView.frame = CGRectMake(labelsBounds.origin.x,
labelsBounds.origin.y + topLabelHeight + kLabelVSpacing,
labelsBounds.size.width,
kAudioProgressViewHeight);
bottomLabel.frame = CGRectMake(labelsBounds.origin.x,
labelsBounds.origin.y + topLabelHeight + kAudioProgressViewHeight + kLabelVSpacing * 2,
labelsBounds.size.width,
bottomLabelHeight);
if (!self.incoming) {
self.attachmentUploadView = [[AttachmentUploadView alloc] initWithAttachment:self.attachment
superview:mediaView
attachmentStateCallback:nil];
}
return mediaView;
}
- (void)ensureAudioDurationSeconds
{
if (self.audioDurationSeconds == 0.f) {
NSError *error;
AVAudioPlayer *audioPlayer = [[AVAudioPlayer alloc] initWithContentsOfURL:self.fileURL error:&error];
OWSAssert(!error);
if (!error) {
self.audioDurationSeconds = (CGFloat)[audioPlayer duration];
}
}
}
- (CGSize)mediaViewDisplaySize {
CGSize size = [super mediaViewDisplaySize];
if ([self isAudio]) {
size.height = (CGFloat)ceil(self.audioBubbleHeight + self.vMargin * 2);
} else if ([self isVideo]) {
return [self ows_adjustBubbleSize:size forImage:self.image];
}
return size;
}
- (UIView *)mediaPlaceholderView {
return [self mediaView];
}
- (NSUInteger)hash {
return [super hash];
}
#pragma mark - OWSMessageEditing Protocol
- (BOOL)canPerformEditingAction:(SEL)action
{
if ([self isVideo]) {
return (action == @selector(copy:) || action == NSSelectorFromString(@"save:"));
} else if ([self isAudio]) {
return (action == @selector(copy:));
}
NSString *actionString = NSStringFromSelector(action);
DDLogError(
@"Unexpected action: %@ for VideoAttachmentAdapter with contentType: %@", actionString, self.contentType);
return NO;
}
- (void)performEditingAction:(SEL)action
{
if ([self isVideo]) {
if (action == @selector(copy:)) {
2017-05-08 17:13:51 +02:00
NSString *utiType = [MIMETypeUtil utiTypeForMIMEType:_contentType];
if (!utiType) {
OWSAssert(0);
utiType = (NSString *)kUTTypeVideo;
}
NSData *data = [NSData dataWithContentsOfURL:self.fileURL];
2017-05-08 17:13:51 +02:00
[UIPasteboard.generalPasteboard setData:data forPasteboardType:utiType];
return;
} else if (action == NSSelectorFromString(@"save:")) {
if (UIVideoAtPathIsCompatibleWithSavedPhotosAlbum(self.fileURL.path)) {
UISaveVideoAtPathToSavedPhotosAlbum(self.fileURL.path, self, nil, nil);
} else {
DDLogWarn(@"cowardly refusing to save incompatible video attachment");
}
}
} else if ([self isAudio]) {
if (action == @selector(copy:)) {
2017-05-08 17:13:51 +02:00
NSString *utiType = [MIMETypeUtil utiTypeForMIMEType:_contentType];
if (!utiType) {
OWSAssert(0);
utiType = (NSString *)kUTTypeAudio;
}
2017-05-08 17:13:51 +02:00
NSData *data = [NSData dataWithContentsOfURL:self.fileURL];
OWSAssert(data);
[UIPasteboard.generalPasteboard setData:data forPasteboardType:utiType];
}
} else {
// Shouldn't get here, as only supported actions should be exposed via canPerformEditingAction
NSString *actionString = NSStringFromSelector(action);
DDLogError(
@"Unexpected action: %@ for VideoAttachmentAdapter with contentType: %@", actionString, self.contentType);
2017-04-11 18:14:26 +02:00
OWSAssert(NO);
}
}
#pragma mark - AudioProgressViewDelegate
- (void)AudioProgressViewWasScrubbedWithProgress:(CGFloat)progress
{
// TODO:
}
#pragma mark - OWSMessageMediaAdapter
- (void)setCellVisible:(BOOL)isVisible
{
// Ignore.
}
2017-04-21 20:19:30 +02:00
- (void)clearCachedMediaViewsIfLastPresentingCell:(id)cell
{
OWSAssert(cell);
2017-04-21 20:19:30 +02:00
if (cell == self.lastPresentingCell) {
[self clearCachedMediaViews];
}
}
@end
2017-03-29 20:49:26 +02:00
NS_ASSUME_NONNULL_END