Closed GoogleCodeExporter closed 8 years ago
I think there is something wrong with libx264.so which I have compiled, now
I am trying to recompile the libx264.so, if I succeed, I would shared my
experience.
Original comment by zhangqia...@gmail.com
on 17 Apr 2012 at 2:48
Samuel, I compiled x264 again, but it still seemed not OK. and this is the
details of how I compiled and linked x264:
1. download x264:
ftp://ftp.videolan.org/pub/x264/snapshots/last_x264.tar.bz2, and tar in
/home/xxx/
2. cd x264/, and type :
/configure --prefix=.. \
--cross-prefix=arm-linux-androideabi- \
--enable-shared \
--enable-pic \
--host=arm-linux
and:
make
then:
make install
then you will find libx264.so and libx264.so.122 in lib/, x264.h and
x264_config.h in include/:
put libx264.so and libx264.so.122 into
android-ndk-r5/platforms/android-8/arch-arm/usr/lib
put x264.h and x264_config.h
into android-ndk-r5/platforms/android-8/arch-arm/usr/include
android ffmpeg ./configure add:
PREBUILT=/home/liang/android-ndk-r5/toolchains/arm-eabi-4.4.0/prebuilt/linux-x86
PLATFORM=/home/liang/android-ndk-r5/platforms/android-8/arch-arm
--enable-libx264
add "-I$PLATFORM/usr/include" to --extra-cflags
add "-rpath-link=$PLATFORM/usr/lib -L$PLATFORM/usr/lib" to --extra-ldflags
and next step is to compile ffmpeg, I believe it would be same as
others said online.
Original comment by zhangqia...@gmail.com
on 17 Apr 2012 at 8:15
What is not OK exactly? You're encoded FLV file contains visible video and
audio data, as one would expect... ?
Original comment by samuel.a...@gmail.com
on 17 Apr 2012 at 8:47
the FLV file contains video and audio data, see the attach file, maybe we
should set the preset of x264, or I compile it again as your method.
Original comment by zhangqia...@gmail.com
on 17 Apr 2012 at 9:01
hi, Samuel, you said "built the included libraries by first applying the
included `ffmpeg-android-20120123.patch` to FFmpeg 0.7.11", I have never
used Fedora before, May I ask how to do it? thanks.
Original comment by zhangqia...@gmail.com
on 18 Apr 2012 at 6:39
This is just for reference. It should work just fine on any Linux distribution.
So, what exactly is wrong with the FLV file you attached? I don't see any major
problem, except that some frames are missing...
Original comment by samuel.a...@gmail.com
on 18 Apr 2012 at 6:42
yes, that's the problem, and the audio seems a little fast, now it confuses me,
and I don't know I do solve it.
Original comment by zhangqia...@gmail.com
on 18 Apr 2012 at 6:47
Is it possible that I lost something configure during compiling?
Original comment by zhangqia...@gmail.com
on 18 Apr 2012 at 6:51
No, like I said, I think it's just your CPU that is too slow to encode in real
time... ARM CPUs are (currently) much slower than x86 CPUs. I think you managed
to compile everything just fine.
Original comment by samuel.a...@gmail.com
on 18 Apr 2012 at 6:54
wow, it does be a tough problem now, is it possible about the the preset of
x264?(forgive me asking the simple question, cause I am a rookie of java,
android and stream media)
Original comment by zhangqia...@gmail.com
on 18 Apr 2012 at 7:06
Of course there are settings to encode faster, with lesser quality, etc. There
are a lot of settings, so take the time to read x264's documentation, blogs,
etc.
Original comment by samuel.a...@gmail.com
on 18 Apr 2012 at 7:09
hi, Samuel,
I set:
video_pts = (double) video_st.pts().val() * video_st.time_base().num() /
video_st.time_base().den();
Log.d(LOG_TAG, "video_st.pts()" + video_st.pts());
and Logcat print out"video_pts: -9.223372036854776E15", and it is same all
the time. so, what should video_c.time_base set with CODE_ID_H264?
Original comment by zhangqia...@gmail.com
on 19 Apr 2012 at 2:03
Hi, Samuel,
video_outbuf_size = imageWidth * imageHeight * 4; // ??
what is it on the basis of?
Original comment by zhangqia...@gmail.com
on 20 Apr 2012 at 2:30
I have no idea what should be done to support x264 properly, so if you figure
it out, please be sure to let me know!
As for video_outbuf_size, I just figured that a compressed frame would not get
bigger than an uncompressed 8-bit RGBA image, the largest pixel format
supported... What should it be?
Original comment by samuel.a...@gmail.com
on 22 Apr 2012 at 2:30
Now I can encode flv(h264/aac), but the audio was not quite right, see attached
file. Maybe you can list out your compiling method, and I compare it with mine.
And I can change some configure or Makefile to match javacv, do you think this
way could help?
Original comment by zhangqia...@gmail.com
on 23 Apr 2012 at 2:32
Attachments:
The configuration I use is all inside this package:
http://code.google.com/p/javacv/downloads/detail?name=ffmpeg-0.7.11-android-arm.
zip
Original comment by samuel.a...@gmail.com
on 23 Apr 2012 at 12:54
Reopening as we have not yet properly handled H.264 and audio, either on
Android or on any other platform...
Original comment by samuel.a...@gmail.com
on 2 Jun 2012 at 5:00
Issue 204 has been merged into this issue.
Original comment by samuel.a...@gmail.com
on 2 Jun 2012 at 5:01
Hello!
What have you done to be able to record h264 video? Do you have an example or
something?
Thanks!
Original comment by Jkolo...@gmail.com
on 2 Jun 2012 at 5:10
This looks like a solution
http://libav-users.943685.n4.nabble.com/libx264-xxx-non-strictly-monotonic-PTS-t
d3275701.html
But someone should try to implement it in javacv.
Original comment by Jkolo...@gmail.com
on 2 Jun 2012 at 6:06
Doesn't seem to work with H.263 either. This doesn't produce a playable file:
IplImage image = cvLoadImage("lena.jpg");
FFmpegFrameRecorder recorder = new FFmpegFrameRecorder("lena.3gp", 512, 512);
recorder.setCodecID(CODEC_ID_H263);
recorder.setFormat("3gp");
recorder.setPixelFormat(PIX_FMT_YUV420P);
recorder.start();
for (int i = 0; i < 100; i++) {
recorder.record(image);
}
recorder.stop();
If anyone knows the solution, please let us know.
Original comment by samuel.a...@gmail.com
on 4 Jun 2012 at 12:34
Hello, thanks for your great work guys! I would like to know about audio merge
in the video. Someone got some advance?
Original comment by igor...@gmail.com
on 6 Jun 2012 at 8:35
It could work with H.264, and you should compile x264 into ffmpeg first, and
the method as mention in Comment 52.
Original comment by zhangqia...@gmail.com
on 11 Jun 2012 at 2:25
Without compiling x264 into ffmpeg, it also works with h.263, and the CodecID
should be CODEC_ID_FLV1
Original comment by zhangqia...@gmail.com
on 11 Jun 2012 at 2:27
Hi Guys!,
I'm trying to do a image + audio video encode. I've downloaded the "fujiaoji" sources (Comment 39) for audio supports. But my application crash, below the log after call start method:
06-11 11:25:16.465: I/Compress(7185): start begin
06-11 11:25:16.490: I/Compress(7185): video start finished
06-11 11:25:16.490: I/Compress(7185): audio start finished
06-11 11:25:16.510: I/Compress(7185): find video encoder
06-11 11:25:16.510: I/Compress(7185): encoder not find
Part of code:
recorder.setCodecVideoID(avcodec.CODEC_ID_FLV1);
recorder.setCodecAudioID(avcodec.CODEC_ID_AAC);
recorder.setFormat("flv");
recorder.setPixelFormat(avutil.PIX_FMT_YUV420P);
recorder.setFrameRate(20);
recorder.start();
Thanks,
Marcelo Alves
Original comment by malves.i...@gmail.com
on 11 Jun 2012 at 2:45
Hi Samuel,
Regarding to the "video_outbuf_size = imageWidth*imageHeight*4;" in the
FFmpegFrameRecorder. I think there is actually a lower limit (~16400) for this
buffer size. If I try to encode a video with dimension < 4100 pixel^2 (w x h),
then it always complains about "buffer smaller than minimum size" (I am using
mpeg4 codec).
If I tweak this video_outbuf_size to imageWidth*imageHeight*10 (or *100) , then
this video lower dimension limit no longer exist.
I suggest to add a lower limit for this "video_outbuf_size" just in case that
imageWidth*imageHeight is too small and the buffer allocated is not reaching
the minimum requirement for encoding.
Original comment by Qzts...@gmail.com
on 11 Jun 2012 at 2:46
I think the real problem happens into FFmpegFrameRecorder line 321.
Message error: 06-11 15:04:55.180: A/libc(11826): Fatal signal 11 (SIGSEGV) at
0xfffffffd (code=1)
Original comment by malves.i...@gmail.com
on 11 Jun 2012 at 6:11
@Qztseng So, what is that lower limit? I can put 10, but it doesn't mean it
will always work either
Original comment by samuel.a...@gmail.com
on 12 Jun 2012 at 1:09
@Samuel
I think the lower limit is around 16400. I currently modified it to check if
the width x height x 4 > than 20000, if not, use 20000 instead of w x h x 4.
Original comment by Qzts...@gmail.com
on 13 Jun 2012 at 10:47
@Jkolobok I think I fixed H.264 encoding, please try the code from the
repository:
http://javacv.googlecode.com/git/src/main/java/com/googlecode/javacv/FrameRecord
er.java
@Qztseng Thanks, I'll be updating that too
Original comment by samuel.a...@gmail.com
on 17 Jun 2012 at 3:28
Oops, that would be:
http://javacv.googlecode.com/git/src/main/java/com/googlecode/javacv/FFmpegFrame
Recorder.java
Original comment by samuel.a...@gmail.com
on 17 Jun 2012 at 3:29
It looks like H.263 has been working all along, but apparently it only supports
a limited set of resolution (128 x 96, 176 x 144, 352 x 288, 704 x 576, and
1408 x 1152), so I hardcoded that. I also included better defaults for common
3gp and mp4 formats such that calls like these do not fail by default anymore:
new FFmpegFrameRecorder("file.3gp", width, height).start();
new FFmpegFrameRecorder("file.mp4", width, height).start();
So, about audio, if anyone wants to contribute working audio support for
FFmpegFrameRecorder, please post your patches here, thanks!
Original comment by samuel.a...@gmail.com
on 17 Jun 2012 at 12:12
Here is my code, encoding both video and audio:
VIDEO_CODE_ID:CODEC_ID_FLV1;
AUDIO_CODE_ID:CODEC_ID_ACC;
Original comment by zhangqia...@gmail.com
on 20 Jun 2012 at 2:27
Attachments:
Great, thanks! If it works, I'll put it in. BTW, have you fixed your issue with
libx264? There's a couple of things I figured out were required by libx264 to
work properly, so it should now work alright, even on Android.
Original comment by samuel.a...@gmail.com
on 20 Jun 2012 at 2:35
Yes, I did. But first, we should compile x246 into ffmpeg, here is my code to
encode h.264/aac:
Original comment by zhangqia...@gmail.com
on 20 Jun 2012 at 2:40
Attachments:
Would you also be able to make a version that minimizes the changes to the
source code, without adding anything that runs only on Android
(android.util.Log at least), but also on Java SE? thank you
Original comment by samuel.a...@gmail.com
on 20 Jun 2012 at 12:34
Ok, I've added audio support to `FFmpegFrameRecorder` (`setAudioChannels(int)`
for int > 0 and `record(Buffer)` alongside `record(IplImage)`)! Please try
these new source files:
http://javacv.googlecode.com/git/src/main/java/com/googlecode/javacv/FrameRecord
er.java
http://javacv.googlecode.com/git/src/main/java/com/googlecode/javacv/FFmpegFrame
Recorder.java
And let me know if you encounter any problems, thanks!
Original comment by samuel.a...@gmail.com
on 25 Jun 2012 at 2:41
Re @87
How do you test audio? Is there some sample code?
Should we call recorder.record(image) and rec.record(buffer) separaterly?
How do you get create buffer with Audio data?
Original comment by marko.ko...@gmail.com
on 25 Jun 2012 at 2:55
Re @87
We should compare video_st to audio_st, for deciding when to write video data
and when to write audio data, to insure the synchronization of video and audio.
Original comment by zhangqia...@gmail.com
on 26 Jun 2012 at 1:34
@89
it should be audio_pts and video_pts
Original comment by zhangqia...@gmail.com
on 26 Jun 2012 at 1:35
Marco, yes, that's what I said. More explicitly, something like:
FrameRecorder recorder = new FFmpegFrameRecorder("filename.avi", width, height, audioChannels);
recorder.start();
// in a loop somewhere
recorder.record(image);
recorder.record(audioSamples);
As for the java.nio.Buffer class, it's all here:
http://docs.oracle.com/javase/6/docs/api/java/nio/Buffer.html
Zhang, no we don't need to, av_interleaved_write_frame() takes care of that:
http://ffmpeg.org/doxygen/0.6/avformat_8h.html#37352ed2c63493c38219d935e71db6c1
But obviously this only works if we actually send `frameRate` frames per
second...
Original comment by samuel.a...@gmail.com
on 26 Jun 2012 at 2:43
Ah, here's some sample code, copy/paste the example from this tutorial:
http://docs.oracle.com/javase/tutorial/sound/converters.html
And replace
// Here, do something useful with the audio data that's
// now in the audioBytes array...
by
recorder.record(someImage); // called according to the frameRate
recorder.record(ByteBuffer.wrap(audioBytes, 0, numBytesRead));
Original comment by samuel.a...@gmail.com
on 26 Jun 2012 at 10:17
Ok, I've also added audio support to FFmpegFrameGrabber, so we can now use them
together to test them out, or to use them as a poor man's transcoder, with
something like this:
FrameGrabber grabber = new FFmpegFrameGrabber("input.avi");
grabber.start();
FrameRecorder recorder = new FFmpegFrameRecorder("output.mp4", grabber.getImageWidth(), grabber.getImageHeight(), grabber.getAudioChannels());
recorder.setFrameRate(grabber.getFrameRate());
recorder.setSampleFormat(grabber.getSampleFormat());
recorder.setSampleRate(grabber.getSampleRate());
recorder.start();
Frame frame;
while ((frame = grabber.grabFrame()) != null) {
recorder.record(frame);
}
recorder.stop();
grabber.stop();
We need these updated classes to try it out:
http://javacv.googlecode.com/git/src/main/java/com/googlecode/javacv/Frame.java
http://javacv.googlecode.com/git/src/main/java/com/googlecode/javacv/FrameGrabbe
r.java
http://javacv.googlecode.com/git/src/main/java/com/googlecode/javacv/FrameRecord
er.java
http://javacv.googlecode.com/git/src/main/java/com/googlecode/javacv/FFmpegFrame
Grabber.java
http://javacv.googlecode.com/git/src/main/java/com/googlecode/javacv/FFmpegFrame
Recorder.java
As usual, let me know of any issues! thanks
Original comment by samuel.a...@gmail.com
on 28 Jun 2012 at 1:49
Actually, I'm trying to test audio on Android, but for some reason can't make
AudioManager to be properly initialized. Whatever combination of source, sample
rate, format and buffer size I try, it fails to initialize, in both emulator
and phone. Has anyone tryed that?
int audioSource = MediaRecorder.AudioSource.DEFAULT;
int[] sampleRates = {44100, 22050, 11025, 8000};
// int sampleRateInHz = 44100;
int channelConfig = AudioFormat.CHANNEL_IN_MONO;
int audioFormat = AudioFormat. ENCODING_PCM_16BIT;
int i = 0;
do {
int sampleRateInHz = sampleRates[i];
bufferSizeInBytes = AudioRecord.getMinBufferSize(sampleRateInHz, channelConfig, audioFormat);
if (bufferSizeInBytes < 1) bufferSizeInBytes = 65536;
audioBuffer = new byte[bufferSizeInBytes];
audioRecord = new AudioRecord(audioSource, sampleRateInHz, channelConfig, audioFormat, bufferSizeInBytes);
} while ((++i<sampleRates.length) & !( audioRecord.getState() == AudioRecord.STATE_INITIALIZED));
Original comment by marko.ko...@gmail.com
on 2 Jul 2012 at 8:46
Here is my recording method, and it works well:
public class CameraView extends SurfaceView implements SurfaceHolder.Callback,
PreviewCallback {
private static final String LOG_TAG = "CameraView & AudioView";
private SurfaceHolder mHolder;
private Camera mCamera;
private boolean isPreviewOn = false;
/* video data */
private int imageWidth;
private int imageHeight;
private int frameRate;
private int sampleVideoBitRate;
private IplImage yuvIplimage;
/* audio data */
private boolean isAudioRecording = false;
private AudioRecord audioRecord;
private Thread audioView;
private int sampleAudioRateInHz;
private int sampleAudioBitRate;
/* recorder */
private volatile boolean isRecorderStart = false;
private FFmpegRecorder recorder;
private String link;
/**
* setup the CameraView and the AudioThread, and initalize the recorder
* @param context android.content.Context
* @param camera hardware.Camera
* @param width width of Camera's size
* @param height height of Camera's size
* @param frame frame rate of camera preview, and for ffmpeg encoding
* @param videoRate video bit rate for ffmpeg encoding
* @param audioRateInHz audio rate for sample and ffmpeg encoding
* @param audioBitRate audio bit rate for sample and ffmpeg encoding
* @param url ffmpeg server connecting link
*/
public CameraView(Context context, Camera camera, int width, int height, int frame, int videoRate,
int audioRateInHz, int audioBitRate, String url) {
super(context);
this.mCamera = camera;
this.imageWidth = width;
this.imageHeight = height;
this.frameRate = frame;
this.sampleVideoBitRate = videoRate;
this.sampleAudioRateInHz = audioRateInHz;
this.sampleAudioBitRate = audioBitRate;
this.link = url;
mHolder = getHolder();
mHolder.addCallback(CameraView.this);
mHolder.setType(SurfaceHolder.SURFACE_TYPE_PUSH_BUFFERS);
mCamera.setPreviewCallback(CameraView.this);
if (Register.isRegistered(context)) {
recorder = new FFmpegRecorder(link, imageWidth, imageHeight, sampleVideoBitRate,
frameRate, sampleAudioRateInHz, sampleAudioBitRate);
audioView = new Thread(new AudioRecordThread());
audioView.start();
} else {
System.exit(0);
}
}
@Override
public void surfaceCreated(SurfaceHolder holder) {
try {
stopPreview();
mCamera.setPreviewDisplay(holder);
} catch (IOException exception) {
mCamera.release();
mCamera = null;
}
}
@Override
public void surfaceChanged(SurfaceHolder holder, int format, int width, int height) {
Camera.Parameters camParams = mCamera.getParameters();
camParams.setPreviewSize(imageWidth, imageHeight);
camParams.setPreviewFrameRate(frameRate);
mCamera.setParameters(camParams);
startPreview();
}
@Override
public void surfaceDestroyed(SurfaceHolder holder) {
mHolder.addCallback(null);
mCamera.setPreviewCallback(null);
isAudioRecording = false;
}
public synchronized void startPreview() {
if (!isPreviewOn && mCamera != null) {
isPreviewOn = true;
mCamera.startPreview();
}
}
public synchronized void stopPreview() {
if (isPreviewOn && mCamera != null) {
isPreviewOn = false;
mCamera.stopPreview();
}
}
@Override
public synchronized void onPreviewFrame(byte[] data, Camera camera) {
try {
if (yuvIplimage == null) {
yuvIplimage = IplImage.create(imageWidth, imageHeight, IPL_DEPTH_8U, 2);
Log.i(LOG_TAG, "create yuvIplimage");
}
/* get video data, yuvIplimage put data */
if (yuvIplimage != null && isRecorderStart) {
yuvIplimage.getByteBuffer().put(data);
}
} catch (Exception e) {
}
}
/**
* thread for getting audio data
* @author Zhang Qianliang
*
*/
public class AudioRecordThread implements Runnable {
@Override
public void run() {
int bufferLength = 0;
int bufferSize;
short[] audioData;
int bufferReadResult;
try {
bufferSize = AudioRecord.getMinBufferSize(sampleAudioRateInHz,
AudioFormat.CHANNEL_CONFIGURATION_MONO, AudioFormat.ENCODING_PCM_16BIT);
if (bufferSize <= 2048) {
bufferLength = 2048;
} else if (bufferSize <= 4096) {
bufferLength = 4096;
}
/* set audio recorder parameters, and start recording */
audioRecord = new AudioRecord(MediaRecorder.AudioSource.MIC, sampleAudioRateInHz,
AudioFormat.CHANNEL_CONFIGURATION_MONO, AudioFormat.ENCODING_PCM_16BIT, bufferLength);
audioData = new short[bufferLength];
audioRecord.startRecording();
Log.d(LOG_TAG, "audioRecord.startRecording()");
isAudioRecording = true;
/* ffmpeg_audio encoding loop */
while (isAudioRecording) {
bufferReadResult = audioRecord.read(audioData, 0, audioData.length);
/* recorder.record(): bufferReadResult == 1024 */
if (bufferReadResult == 1024 && isRecorderStart) {
short[] realAudioData1024 = new short[bufferReadResult];
System.arraycopy(audioData, 0, realAudioData1024, 0, bufferReadResult);
recorder.recording(yuvIplimage, realAudioData1024);
}
/* recorder.record(): bufferReadResult == 2048 */
else if (bufferReadResult == 2048 && isRecorderStart) {
short[] realAudioData2048_1 = new short[1024];
short[] realAudioData2048_2 = new short[1024];
System.arraycopy(audioData, 0, realAudioData2048_1, 0, 1024);
System.arraycopy(audioData, 1024, realAudioData2048_2, 0, 1024);
for (int i = 0; i < 2; i++) {
if (i == 0) {
recorder.recording(yuvIplimage, realAudioData2048_1);
} else if (i == 1) {
recorder.recording(yuvIplimage, realAudioData2048_2);
}
}
}
}
/* encoding finish, release recorder */
if (audioRecord != null) {
try {
audioRecord.stop();
audioRecord.release();
} catch (Exception e) {
e.printStackTrace();
}
audioRecord = null;
}
if (recorder != null && isRecorderStart) {
try {
stop();
} catch (Exception e) {
e.printStackTrace();
}
recorder = null;
}
} catch (Exception e) {
Log.e(LOG_TAG, "get audio data failed");
}
}
}
/**
* judge if recorder started or not
* @return true--> started, false--> not started
*/
public boolean isRecorderStart() {
return isRecorderStart;
}
/**
* to start or stop recording
* @param status ture-->to start, false-->to stop
*/
public void setRecorderStatus(boolean status) {
if (!this.isRecorderStart) {
start();
}
this.isRecorderStart = status;
}
/**
* start recorder
*/
private void start() {
recorder.start();
}
/**
* stop recorder
*/
private void stop() {
recorder.stop();
recorder.release();
}
}
Original comment by zhangqia...@gmail.com
on 3 Jul 2012 at 1:57
In this example, where is FFmpegRecorder defined?
Original comment by mik...@gmail.com
on 3 Jul 2012 at 6:19
@96
/* recorder */
private volatile boolean isRecorderStart = false;
private FFmpegRecorder recorder;
private String link;
Original comment by yonyf...@gmail.com
on 9 Jul 2012 at 3:54
JavaCV 0.2 now comes with all relevant changes to support H.264 and audio with
FFmpeg 0.11.1. Let me know that it works well, thanks!
Original comment by samuel.a...@gmail.com
on 22 Jul 2012 at 5:37
think he meant where u defined that class as is not part of javacv. Can you
post ur FFmpegRecorder class please? or full project for the camera stuff will
be good ;)
Original comment by florin.m...@gmail.com
on 22 Jul 2012 at 7:04
a demo to record stream(CODEC_ID_FLV1/CODEC_ID_AAC):
Original comment by zhangqia...@gmail.com
on 24 Jul 2012 at 9:15
Attachments:
Original issue reported on code.google.com by
zhangqia...@gmail.com
on 27 Feb 2012 at 3:22