Results 1 to 6 of 6

Thread: Unable to record video Stream

  1. #1
    Join Date
    Jul 2014
    Posts
    13

    Default Unable to record video Stream

    I'm able to view my video streams on Wowza -> Live Incoming Streams and they are also playable on Live Test Players. Following is the network progress while I'm streaming video.

    Stream Uptime
    myStream up since 30 Sep 2014 03:26:46 PM
    myStream up for about a few seconds
    Network Throughput
    Bytes In: 826073  @  371.490 Kbits/s
    Bytes Out: 0  @  0.000 bits/s
    When I record that stream a file is created, but when I play it video/audio is not displayed. I've also tried VOD application but still unable to get proper recorded video file.

    Can you please suggest what's going wrong?

  2. #2

    Default

    Hello there.

    What method are you attempting to record with? Have you followed one of these guides:
    How to record live streams (HTTPLiveStreamRecord)
    How to record live streams (Wowza Streaming Engine)

    What is the source of the live stream? And the video and audio codec?

    Can you start Wowza in stand-alone mode [install-dir]/bin/startup.bat|startup.sh and post the output from the console showing playback of the recorded stream?

    Thank you.

    Salvadore

  3. #3
    Join Date
    Jul 2014
    Posts
    13

    Default

    I'm developing video streaming application for Android. I've used FFmpeg Encoder and JavaCV Library for this purpose. Audio is being recorded but video is not recording.

    And following error is shown when I try to launch Wowza in stand-alone mode:-
    BindException: Address already in use: bind|at sun.nio.ch.Net.bind0(Native Metho
    d)|at sun.nio.ch.Net.bind(Net.java:414)|at sun.nio.ch.Net.bind(Net.java:406)|at
    sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:214)|at sun
    .nio.ch.ServerSocketAdaptor.bind(ServerSocketAdaptor.java:74)
    My java class is appended below:-

    package com.example.javacv.stream.test2;
    
    import static com.googlecode.javacv.cpp.opencv_core.IPL_DEPTH_8U;
    import java.io.IOException;
    import java.nio.ShortBuffer;
    import android.app.Activity;
    import android.content.Context;
    import android.content.Intent;
    import android.content.pm.ActivityInfo;
    import android.graphics.Bitmap;
    import android.graphics.Canvas;
    import android.hardware.Camera;
    import android.hardware.Camera.PreviewCallback;
    import android.media.AudioFormat;
    import android.media.AudioRecord;
    import android.media.MediaRecorder;
    import android.os.Bundle;
    import android.os.PowerManager;
    import android.util.Log;
    import android.view.Display;
    import android.view.KeyEvent;
    import android.view.SurfaceHolder;
    import android.view.SurfaceView;
    import android.view.View;
    import android.view.View.OnClickListener;
    import android.widget.LinearLayout;
    import android.widget.Toast;
    
    import com.googlecode.javacv.FFmpegFrameRecorder;
    import com.googlecode.javacv.cpp.opencv_core.IplImage;
    
    public class MainActivity extends Activity {
    
    	private final static String LOG_TAG = "MainActivity";
    	private PowerManager.WakeLock mWakeLock;
    
            //WOWZA SERVER IP ADDRESS IS REMOVED FOR SECURITY PURPOSE
    	private String ffmpeg_link = "rtmp://live:live@WOWZA_IP_ADDRESS:1935/live/myStream";
    
    	private volatile FFmpegFrameRecorder recorder;
    	boolean recording = false;
    	long startTime = 0;
    	private int sampleAudioRateInHz = 44100;
    	private int imageWidth = 100;
    	private int imageHeight = 100;
    	private int frameRate = 24;
    	private Thread audioThread;
    	volatile boolean runAudioThread = true;
    	private AudioRecord audioRecord;
    	private AudioRecordRunnable audioRecordRunnable;
    	private CameraView cameraView;
    	private IplImage yuvIplimage = null;
    	private LinearLayout mainLayout;
    
    	@Override
    	public void onCreate(Bundle savedInstanceState) {
    		super.onCreate(savedInstanceState);
    		//setRequestedOrientation(ActivityInfo.SCREEN_ORIENTATION_LANDSCAPE);
    		setContentView(R.layout.activity_main);
    		initLayout();
    		initRecorder();
    		startRecording();
    	}
    
    	@Override
    	protected void onResume() {
    		super.onResume();
    
    		if (mWakeLock == null) {
    			PowerManager pm = (PowerManager) getSystemService(Context.POWER_SERVICE);
    			mWakeLock = pm.newWakeLock(PowerManager.SCREEN_BRIGHT_WAKE_LOCK,
    					LOG_TAG);
    			mWakeLock.acquire();
    		}
    	}
    
    	@Override
    	protected void onPause() {
    		super.onPause();
    
    		if (mWakeLock != null) {
    			mWakeLock.release();
    			mWakeLock = null;
    		}
    	}
    
    	@Override
    	protected void onDestroy() {
    		super.onDestroy();
    
    		recording = false;
    		stopRecording();
    	}
    
    	private void initLayout() {
    
    		mainLayout = (LinearLayout) this.findViewById(R.id.record_layout);
    
    		Display display = getWindowManager().getDefaultDisplay();
    		int width = display.getWidth();
    		int height = display.getHeight();
    		imageWidth = width;
    		imageHeight = height;
    
    		cameraView = new CameraView(this);
    
    		LinearLayout.LayoutParams layoutParam = new LinearLayout.LayoutParams(
    				imageWidth, imageHeight);
    
    		mainLayout.addView(cameraView, layoutParam);
    
    		Log.v(LOG_TAG, "added cameraView to mainLayout");
    	}
    
    	private void initRecorder() {
    		Log.w(LOG_TAG, "initRecorder");
    
    		if (yuvIplimage == null) {
    			// Recreated after frame size is set in surface change method
    			yuvIplimage = IplImage.create(imageWidth, imageHeight,
    					IPL_DEPTH_8U, 2);
    			// yuvIplimage = IplImage.create(imageWidth, imageHeight,
    			// IPL_DEPTH_32S, 2);
    
    			Log.v(LOG_TAG, "IplImage.create");
    		}
    
    		recorder = new FFmpegFrameRecorder(ffmpeg_link, imageWidth,
    				imageHeight, 1);
    		Log.v(LOG_TAG, "FFmpegFrameRecorder: " + ffmpeg_link + " imageWidth: "
    				+ imageWidth + " imageHeight " + imageHeight);
    
    		recorder.setFormat("flv");
    		Log.v(LOG_TAG, "recorder.setFormat(\"flv\")");
    
    		recorder.setSampleRate(sampleAudioRateInHz);
    		Log.v(LOG_TAG, "recorder.setSampleRate(sampleAudioRateInHz)");
    
    		// re-set in the surface changed method as well
    		recorder.setFrameRate(frameRate);
    
    		Log.v(LOG_TAG, "recorder.setFrameRate(frameRate)");
    
    		// Create audio recording thread
    		audioRecordRunnable = new AudioRecordRunnable();
    		audioThread = new Thread(audioRecordRunnable);
    	}
    
    	// Start the capture
    	public void startRecording() {
    		try {
    			Toast.makeText(getApplicationContext(), "Recording started",
    					Toast.LENGTH_LONG).show();
    			recorder.start();
    			startTime = System.currentTimeMillis();
    			recording = true;
    			audioThread.start();
    		} catch (FFmpegFrameRecorder.Exception e) {
    			Toast.makeText(getApplicationContext(), "Error\n" + e.toString(),
    					Toast.LENGTH_LONG).show();
    			e.printStackTrace();
    		} catch (Exception e) {
    			Toast.makeText(getApplicationContext(), "Error\n" + e.toString(),
    					Toast.LENGTH_LONG).show();
    		}
    	}
    
    	public void stopRecording() {
    		// This should stop the audio thread from running
    		runAudioThread = false;
    
    		if (recorder != null && recording) {
    			recording = false;
    			Log.v(LOG_TAG,
    					"Finishing recording, calling stop and release on recorder");
    			try {
    				recorder.stop();
    				recorder.release();
    			} catch (FFmpegFrameRecorder.Exception e) {
    				e.printStackTrace();
    			}
    			recorder = null;
    		}
    	}
    
    	@Override
    	public boolean onKeyDown(int keyCode, KeyEvent event) {
    		// Quit when back button is pushed
    		if (keyCode == KeyEvent.KEYCODE_BACK) {
    			if (recording) {
    				stopRecording();
    			}
    			finish();
    			return true;
    		}
    		return super.onKeyDown(keyCode, event);
    	}
    
    	/*
    	 * @Override public void onClick(View v) { if (!recording) {
    	 * startRecording(); Log.w(LOG_TAG, "Start Button Pushed");
    	 * recordButton.setText("Stop"); } else { stopRecording(); Log.w(LOG_TAG,
    	 * "Stop Button Pushed"); recordButton.setText("Start"); } }
    	 */
    
    	// ---------------------------------------------
    	// audio thread, gets and encodes audio data
    	// ---------------------------------------------
    	class AudioRecordRunnable implements Runnable {
    
    		@Override
    		public void run() {
    			// Set the thread priority
    			android.os.Process
    					.setThreadPriority(android.os.Process.THREAD_PRIORITY_URGENT_AUDIO);
    
    			// Audio
    			int bufferSize;
    			short[] audioData;
    			int bufferReadResult;
    
    			bufferSize = AudioRecord.getMinBufferSize(sampleAudioRateInHz,
    					AudioFormat.CHANNEL_CONFIGURATION_MONO,
    					AudioFormat.ENCODING_PCM_16BIT);
    			audioRecord = new AudioRecord(MediaRecorder.AudioSource.MIC,
    					sampleAudioRateInHz,
    					AudioFormat.CHANNEL_CONFIGURATION_MONO,
    					AudioFormat.ENCODING_PCM_16BIT, bufferSize);
    
    			audioData = new short[bufferSize];
    
    			Log.d(LOG_TAG, "audioRecord.startRecording()");
    			audioRecord.startRecording();
    
    			// Audio Capture/Encoding Loop
    			while (runAudioThread) {
    				// Read from audioRecord
    				bufferReadResult = audioRecord.read(audioData, 0,
    						audioData.length);
    				if (bufferReadResult > 0) {
    					// Log.v(LOG_TAG,"audioRecord bufferReadResult: " +
    					// bufferReadResult);
    
    					// Changes in this variable may not be picked up despite it
    					// being "volatile"
    					if (recording) {
    						try {
    							// Write to FFmpegFrameRecorder
    							recorder.record(ShortBuffer.wrap(audioData, 0,
    									bufferReadResult));
    						} catch (FFmpegFrameRecorder.Exception e) {
    							Log.v(LOG_TAG, e.getMessage());
    							e.printStackTrace();
    						}
    					}
    				}
    			}
    			Log.v(LOG_TAG, "AudioThread Finished");
    
    			/* Capture/Encoding finished, release recorder */
    			if (audioRecord != null) {
    				audioRecord.stop();
    				audioRecord.release();
    				audioRecord = null;
    				Log.v(LOG_TAG, "audioRecord released");
    			}
    		}
    	}
    
    	class CameraView extends SurfaceView implements SurfaceHolder.Callback,
    			PreviewCallback {
    
    		private boolean previewRunning = false;
    		private SurfaceHolder holder;
    		private Camera camera;
    		private byte[] previewBuffer;
    		long videoTimestamp = 0;
    		Bitmap bitmap;
    		Canvas canvas;
    
    		public CameraView(Context _context) {
    			super(_context);
    
    			holder = this.getHolder();
    			holder.addCallback(this);
    			holder.setType(SurfaceHolder.SURFACE_TYPE_PUSH_BUFFERS);
    		}
    
    		@Override
    		public void surfaceCreated(SurfaceHolder holder) {
    			camera = Camera.open();
    
    			try {
    				camera.setPreviewDisplay(holder);
    				camera.setPreviewCallback(this);
    
    				Camera.Parameters currentParams = camera.getParameters();
    				Log.v(LOG_TAG,
    						"Preview Framerate: "
    								+ currentParams.getPreviewFrameRate());
    				Log.v(LOG_TAG,
    						"Preview imageWidth: "
    								+ currentParams.getPreviewSize().width
    								+ " imageHeight: "
    								+ currentParams.getPreviewSize().height);
    
    				// Use these values
    				imageWidth = currentParams.getPreviewSize().width;
    				imageHeight = currentParams.getPreviewSize().height;
    				frameRate = currentParams.getPreviewFrameRate();
    
    				bitmap = Bitmap.createBitmap(imageWidth, imageHeight,
    						Bitmap.Config.ALPHA_8);
    
    				/*
    				 * Log.v(LOG_TAG,"Creating previewBuffer size: " + imageWidth *
    				 * imageHeight *
    				 * ImageFormat.getBitsPerPixel(currentParams.getPreviewFormat
    				 * ())/8); previewBuffer = new byte[imageWidth * imageHeight *
    				 * ImageFormat
    				 * .getBitsPerPixel(currentParams.getPreviewFormat())/8];
    				 * camera.addCallbackBuffer(previewBuffer);
    				 * camera.setPreviewCallbackWithBuffer(this);
    				 */
    
    				camera.startPreview();
    				previewRunning = true;
    			} catch (IOException e) {
    				Log.v(LOG_TAG, e.getMessage());
    				e.printStackTrace();
    			}
    		}
    
    		public void surfaceChanged(SurfaceHolder holder, int format, int width,
    				int height) {
    			Log.v(LOG_TAG, "Surface Changed: width " + width + " height: "
    					+ height);
    
    			// We would do this if we want to reset the camera parameters
    			/*
    			 * if (!recording) { if (previewRunning){ camera.stopPreview(); }
    			 * 
    			 * try { //Camera.Parameters cameraParameters =
    			 * camera.getParameters(); //p.setPreviewSize(imageWidth,
    			 * imageHeight); //p.setPreviewFrameRate(frameRate);
    			 * //camera.setParameters(cameraParameters);
    			 * 
    			 * camera.setPreviewDisplay(holder); camera.startPreview();
    			 * previewRunning = true; } catch (IOException e) {
    			 * Log.e(LOG_TAG,e.getMessage()); e.printStackTrace(); } }
    			 */
    
    			// Get the current parameters
    			Camera.Parameters currentParams = camera.getParameters();
    			Log.v(LOG_TAG,
    					"Preview Framerate: " + currentParams.getPreviewFrameRate());
    			Log.v(LOG_TAG,
    					"Preview imageWidth: "
    							+ currentParams.getPreviewSize().width
    							+ " imageHeight: "
    							+ currentParams.getPreviewSize().height);
    
    			// Use these values
    			imageWidth = currentParams.getPreviewSize().width;
    			imageHeight = currentParams.getPreviewSize().height;
    			frameRate = currentParams.getPreviewFrameRate();
    
    			// Create the yuvIplimage if needed
    			yuvIplimage = IplImage.create(imageWidth, imageHeight,
    					IPL_DEPTH_8U, 2);
    			// yuvIplimage = IplImage.create(imageWidth, imageHeight,
    			// IPL_DEPTH_32S, 2);
    		}
    
    		@Override
    		public void surfaceDestroyed(SurfaceHolder holder) {
    			try {
    				camera.setPreviewCallback(null);
    				previewRunning = false;
    				camera.release();
    
    			} catch (RuntimeException e) {
    				Log.v(LOG_TAG, e.getMessage());
    				e.printStackTrace();
    			}
    		}
    
    		@Override
    		public void onPreviewFrame(byte[] data, Camera camera) {
    
    			if (yuvIplimage != null && recording) {
    				videoTimestamp = 1000 * (System.currentTimeMillis() - startTime);
    
    				// Put the camera preview frame right into the yuvIplimage
    				// object
    				yuvIplimage.getByteBuffer().put(data);
    
    				// FAQ about IplImage:
    				// - For custom raw processing of data, getByteBuffer() returns
    				// an NIO direct
    				// buffer wrapped around the memory pointed by imageData, and
    				// under Android we can
    				// also use that Buffer with Bitmap.copyPixelsFromBuffer() and
    				// copyPixelsToBuffer().
    				// - To get a BufferedImage from an IplImage, we may call
    				// getBufferedImage().
    				// - The createFrom() factory method can construct an IplImage
    				// from a BufferedImage.
    				// - There are also a few copy*() methods for
    				// BufferedImage<->IplImage data transfers.
    
    				// Let's try it..
    				// This works but only on transparency
    				// Need to find the right Bitmap and IplImage matching types
    
    				/*
    				 * bitmap.copyPixelsFromBuffer(yuvIplimage.getByteBuffer());
    				 * //bitmap.setPixel(10,10,Color.MAGENTA);
    				 * 
    				 * canvas = new Canvas(bitmap); Paint paint = new Paint();
    				 * paint.setColor(Color.GREEN); float leftx = 20; float topy =
    				 * 20; float rightx = 50; float bottomy = 100; RectF rectangle =
    				 * new RectF(leftx,topy,rightx,bottomy);
    				 * canvas.drawRect(rectangle, paint);
    				 * 
    				 * bitmap.copyPixelsToBuffer(yuvIplimage.getByteBuffer());
    				 */
    				// Log.v(LOG_TAG,"Writing Frame");
    
    				try {
    
    					// Get the correct time
    					recorder.setTimestamp(videoTimestamp);
    
    					// Record the image into FFmpegFrameRecorder
    					recorder.record(yuvIplimage);
    
    				} catch (FFmpegFrameRecorder.Exception e) {
    					Log.v(LOG_TAG, e.getMessage());
    					e.printStackTrace();
    				}
    			}
    		}
    	}
    
    }
    Last edited by AhmedFaraz; 09-30-2014 at 11:46 PM.

  4. #4
    Join Date
    Jul 2014
    Posts
    13

    Default

    Using RTSP I'm now able to stream and record my videos. I've used this post.

  5. #5

    Default

    Thank you very much for the update and helpful guide. Glad to hear you got this working.

    Salvadore

  6. #6

    Default

    Thanks AhmedFaraz for useful information.


    ___________________
    Nail Designs

Similar Threads

  1. how to record stream in .mp4 format video using flex.
    By shanth.5577 in forum Tutorials Discussion
    Replies: 6
    Last Post: 11-20-2014, 02:35 PM
  2. Video player and says unable to play video
    By aqeelashraf in forum Live Streaming and Encoder Discussion
    Replies: 3
    Last Post: 01-20-2014, 06:06 AM
  3. How to record video stream to custom file name
    By terbooter in forum General Forum
    Replies: 3
    Last Post: 01-12-2013, 08:55 AM

Posting Permissions

  • You may not post new threads
  • You may not post replies
  • You may not post attachments
  • You may not edit your posts
  •