//-------------------------------------------------------------------------------------------------- BOOL CSoundStream::Decompress(unsigned char *dest) { u32 dwSrcSize = dwSrcBufSize; BOOL r = true; VERIFY (hAcmStream); // check for EOF if (dwDecPos+dwSrcSize>dwTotalSize) { dwSrcSize=dwTotalSize-dwDecPos; r=false; } hf->r (WaveSource,dwSrcSize); stream.cbStruct=sizeof(stream); stream.fdwStatus=0; stream.pbSrc=WaveSource; stream.cbSrcLength=dwSrcSize; stream.pbDst=dest; stream.cbDstLength=dwDestBufSize; CHK_DX(acmStreamPrepareHeader(hAcmStream,&stream,0)); CHK_DX(acmStreamConvert(hAcmStream,&stream,0)); CHK_DX(acmStreamUnprepareHeader(hAcmStream,&stream,0)); dwDecPos+=stream.cbSrcLengthUsed; AppWriteDataToBuffer(writepos,WaveDest,stream.cbDstLengthUsed); return r; }
static DWORD wodPrepare(WAVEMAPDATA* wom, LPWAVEHDR lpWaveHdrSrc, DWORD dwParam2) { PACMSTREAMHEADER ash; DWORD size; DWORD dwRet; LPWAVEHDR lpWaveHdrDst; TRACE("(%p %p %08x)\n", wom, lpWaveHdrSrc, dwParam2); if (!wom->hAcmStream) return waveOutPrepareHeader(wom->u.out.hInnerWave, lpWaveHdrSrc, dwParam2); if (acmStreamSize(wom->hAcmStream, lpWaveHdrSrc->dwBufferLength, &size, ACM_STREAMSIZEF_SOURCE) != MMSYSERR_NOERROR) { WARN("acmStreamSize failed\n"); return MMSYSERR_ERROR; } ash = HeapAlloc(GetProcessHeap(), 0, sizeof(ACMSTREAMHEADER) + sizeof(WAVEHDR) + size); if (ash == NULL) { WARN("no memory\n"); return MMSYSERR_NOMEM; } ash->cbStruct = sizeof(*ash); ash->fdwStatus = 0L; ash->dwUser = (DWORD_PTR)lpWaveHdrSrc; ash->pbSrc = (LPBYTE)lpWaveHdrSrc->lpData; ash->cbSrcLength = lpWaveHdrSrc->dwBufferLength; /* ash->cbSrcLengthUsed */ ash->dwSrcUser = lpWaveHdrSrc->dwUser; /* FIXME ? */ ash->pbDst = (LPBYTE)ash + sizeof(ACMSTREAMHEADER) + sizeof(WAVEHDR); ash->cbDstLength = size; /* ash->cbDstLengthUsed */ ash->dwDstUser = 0; /* FIXME ? */ dwRet = acmStreamPrepareHeader(wom->hAcmStream, ash, 0L); if (dwRet != MMSYSERR_NOERROR) { WARN("acmStreamPrepareHeader failed\n"); goto errCleanUp; } lpWaveHdrDst = (LPWAVEHDR)((LPSTR)ash + sizeof(ACMSTREAMHEADER)); lpWaveHdrDst->lpData = (LPSTR)ash->pbDst; lpWaveHdrDst->dwBufferLength = size; /* conversion is not done yet */ lpWaveHdrDst->dwFlags = 0; lpWaveHdrDst->dwLoops = 0; dwRet = waveOutPrepareHeader(wom->u.out.hInnerWave, lpWaveHdrDst, sizeof(*lpWaveHdrDst)); if (dwRet != MMSYSERR_NOERROR) { WARN("waveOutPrepareHeader failed\n"); goto errCleanUp; } lpWaveHdrSrc->reserved = (DWORD_PTR)ash; lpWaveHdrSrc->dwFlags = WHDR_PREPARED; TRACE("=> (0)\n"); return MMSYSERR_NOERROR; errCleanUp: TRACE("=> (%d)\n", dwRet); HeapFree(GetProcessHeap(), 0, ash); return dwRet; }
int ToADPCM(short *Source, short *Dest, int Size) { int Src_size; int Dest_Size; Wave_Format.wFormatTag = WAVE_FORMAT_PCM; Wave_Format.nChannels = 1; Wave_Format.cbSize = 0; Wave_Format.wBitsPerSample = 16; Wave_Format.nSamplesPerSec = 44100; Wave_Format.nBlockAlign = Wave_Format.nChannels * Wave_Format.wBitsPerSample / 8; Wave_Format.nAvgBytesPerSec = Wave_Format.nSamplesPerSec * Wave_Format.nBlockAlign; ADPCM_Format.wfx.wFormatTag = WAVE_FORMAT_IMA_ADPCM; acmFormatSuggest(NULL, (LPWAVEFORMATEX) &Wave_Format, (LPWAVEFORMATEX) &ADPCM_Format, sizeof(TrueSpeech_Format), ACM_FORMATSUGGESTF_WFORMATTAG); acmStreamOpen(&Pack_Stream, NULL, (LPWAVEFORMATEX) &Wave_Format, (LPWAVEFORMATEX) &ADPCM_Format, NULL, 0, 0, ACM_STREAMOPENF_NONREALTIME); Src_size = Size; unsigned long rawbufsize = 0; acmStreamSize(Pack_Stream, Src_size, &rawbufsize, ACM_STREAMSIZEF_SOURCE); Uint8 *Pack_Buf = (Uint8 *) malloc(Src_size + 8); memset(Pack_Buf, 0, Src_size + 8); Uint8 *rawbuf = (Uint8 *) malloc(rawbufsize + 8); memset(rawbuf, 0, rawbufsize + 8); ACMSTREAMHEADER Pack_Stream_Head; ZeroMemory(&Pack_Stream_Head, sizeof(ACMSTREAMHEADER)); Pack_Stream_Head.cbStruct = sizeof(ACMSTREAMHEADER); Pack_Stream_Head.pbSrc = (Uint8 *) Pack_Buf; Pack_Stream_Head.cbSrcLength = Src_size; Pack_Stream_Head.pbDst = rawbuf; Pack_Stream_Head.cbDstLength = rawbufsize; acmStreamPrepareHeader(Pack_Stream, &Pack_Stream_Head, 0); memcpy(Pack_Buf, Source, Src_size); acmStreamConvert(Pack_Stream, &Pack_Stream_Head, 0); Dest_Size = Pack_Stream_Head.cbDstLengthUsed; if(Dest_Size < Src_size) { memcpy(Dest, rawbuf, Dest_Size); } else { Dest_Size = 0; } acmStreamUnprepareHeader(Pack_Stream, &Pack_Stream_Head, 0); if(rawbuf) free(rawbuf); if(Pack_Buf) free(Pack_Buf); acmStreamClose(Pack_Stream, 0); return(Dest_Size); }
static gboolean acmmp3dec_setup (ACMMP3Dec * dec) { MMRESULT res; int destBufferSize; acmmp3dec_set_input_format (dec); acmmp3dec_set_output_format (dec); res = acmStreamOpen (&dec->stream, NULL, (LPWAVEFORMATEX) & dec->infmt, &dec->outfmt, 0, 0, 0, 0); if (res) { GST_WARNING_OBJECT (dec, "Failed to open ACM stream: %d", res); return FALSE; } dec->header.cbStruct = sizeof (ACMSTREAMHEADER); dec->header.fdwStatus = 0; dec->header.dwUser = 0; dec->header.pbSrc = (BYTE *) g_malloc (ACM_BUFFER_SIZE); dec->header.cbSrcLength = ACM_BUFFER_SIZE; dec->header.cbSrcLengthUsed = 0; dec->header.dwSrcUser = 0; /* Ask what buffer size we need to use for our output */ acmStreamSize (dec->stream, ACM_BUFFER_SIZE, (LPDWORD) & destBufferSize, ACM_STREAMSIZEF_SOURCE); dec->header.pbDst = (BYTE *) g_malloc (destBufferSize); dec->header.cbDstLength = destBufferSize; dec->header.cbDstLengthUsed = 0; dec->header.dwDstUser = 0; res = acmStreamPrepareHeader (dec->stream, &dec->header, 0); if (res) { GST_WARNING_OBJECT (dec, "Failed to prepare ACM stream: %x", res); return FALSE; } dec->output_caps = acmmp3dec_caps_from_format (&dec->outfmt); if (dec->output_caps) { gst_pad_set_caps (dec->srcpad, dec->output_caps); } dec->timestamp = GST_CLOCK_TIME_NONE; dec->is_setup = TRUE; return TRUE; }
LRESULT CAcm::Convert(LPVOID lpDestBuf){ // ここでDirectSoundのLockしたメモリポインタを渡す if (!m_bOpen) { return 1; // Openできてへんのに呼ぶなっちゅーに! } m_acmheader.cbStruct = sizeof(m_acmheader); m_acmheader.pbSrc = (BYTE*)m_lpSrcBuf; m_acmheader.cbSrcLength = m_dwSrcLength; m_acmheader.pbDst = (BYTE*)lpDestBuf; // ここにコピーしたいねん! m_acmheader.cbDstLength = m_dwDestLength; if (acmStreamPrepareHeader(m_hAcm,&m_acmheader,NULL)!=0) { return 2; // 勘弁して〜(笑) } if (acmStreamConvert(m_hAcm,&m_acmheader,NULL)!=0){ return 3; // ダメじゃん(笑) } return 0; // 任務終了 }
/// <summary> /// <c>wACMStreamPrepareHeader</c> /// </summary> /// <remarks> /// </remarks> /// <param name="has"></param> /// <param name="pash"></param> /// <param name="fdwPrepare"></param> /// <returns>HRESULT __stdcall</returns> HRESULT __stdcall wACMStreamPrepareHeader(HACMSTREAM has, PACMSTREAMHEADER pash, ULONG fdwPrepare) { PSTR pszErrorMessage; MMRESULT hResult; InternalFunctionSpew("GameOS_DirectSound", "acmStreamPrepareHeader(0x%x, 0x%x, 0x%x)", has, pash, fdwPrepare); hResult = acmStreamPrepareHeader(has, pash, fdwPrepare); if (MMFAILED(hResult)) { pszErrorMessage = ErrorNumberToMessage(hResult); //PAUSE( if ( InternalFunctionPause( "FAILED (0x%x - %s) - acmStreamPrepareHeader(0x%x, 0x%x, 0x%x)", hResult, pszErrorMessage, has, pash, fdwPrepare) ) ENTER_DEBUGGER; } return hResult; }
void acm_cv_convert (const converter_fmt_t *cfmt, u_char *state, sample *src_buf, int src_len, sample *dst_buf, int dst_len) { ACMSTREAMHEADER ash; LPHACMSTREAM lphs; UNUSED(cfmt); memset(&ash, 0, sizeof(ash)); ash.cbStruct = sizeof(ash); ash.pbSrc = (LPBYTE)src_buf; ash.cbSrcLength = src_len * sizeof(sample); ash.pbDst = (LPBYTE)dst_buf; ash.cbDstLength = dst_len * sizeof(sample); lphs = (LPHACMSTREAM)state; if (acmStreamPrepareHeader(*lphs, &ash, 0) || acmStreamConvert(*lphs, &ash, ACM_STREAMCONVERTF_BLOCKALIGN)) { memset(dst_buf, 0, dst_len * sizeof(sample)); } return; }
void CACMStream::Prepare(LPVOID src, DWORD size) { m_hdr->pbSrc = (LPBYTE) src; m_hdr->cbSrcLength = m_hdr->dwSrcUser = size; // get the suggested output buffer size DWORD dstSize; m_mmr = acmStreamSize(m_strm, size, &dstSize, ACM_STREAMSIZEF_SOURCE); if (m_mmr) raiseError(m_mmr, "Error evaluating destination buffer size"); m_hdr->pbDst = (LPBYTE) GlobalAlloc(GMEM_FIXED, dstSize); if (!m_hdr->pbDst) { m_mmr = MMSYSERR_NOMEM; raiseError(m_mmr, "Error allocating destination buffer"); } m_hdr->cbDstLength = m_hdr->dwDstUser = dstSize; m_mmr = acmStreamPrepareHeader(m_strm, m_hdr, 0); if (m_mmr) raiseError(m_mmr, "Error prparing stream header"); }
int ToMP3(short *Source, short *Dest, int Size, int BitRate) { int Src_size; int Dest_Size; Wave_Format.wFormatTag = WAVE_FORMAT_PCM; Wave_Format.nChannels = 1; Wave_Format.cbSize = 0; Wave_Format.wBitsPerSample = 16; Wave_Format.nSamplesPerSec = 44100; Wave_Format.nBlockAlign = Wave_Format.nChannels * Wave_Format.wBitsPerSample / 8; Wave_Format.nAvgBytesPerSec = Wave_Format.nSamplesPerSec * Wave_Format.nBlockAlign; MP3_Format.wfx.wFormatTag = WAVE_FORMAT_MPEGLAYER3; MP3_Format.wfx.cbSize = MPEGLAYER3_WFX_EXTRA_BYTES; MP3_Format.wfx.nChannels = 1; MP3_Format.wfx.nSamplesPerSec = 44100; MP3_Format.wfx.nAvgBytesPerSec = BitRate * (1000 / 8); MP3_Format.wfx.wBitsPerSample = 0; MP3_Format.wfx.nBlockAlign = 1; MP3_Format.wID = MPEGLAYER3_ID_MPEG; MP3_Format.fdwFlags = MPEGLAYER3_FLAG_PADDING_OFF; MP3_Format.nBlockSize = 0; MP3_Format.nFramesPerBlock = 0; MP3_Format.nCodecDelay = 0; acmStreamOpen(&Pack_Stream, NULL, (LPWAVEFORMATEX) &Wave_Format, (LPWAVEFORMATEX) &MP3_Format, NULL, 0, 0, 0); Src_size = Size; unsigned long rawbufsize = 0; acmStreamSize(Pack_Stream, Src_size, &rawbufsize, ACM_STREAMSIZEF_SOURCE); rawbufsize += MP3_FRAMES_LAG * 2; Uint8 *Pack_Buf = (Uint8 *) malloc(Src_size + (MP3_FRAMES_LAG * 4) + 8); memset(Pack_Buf, 0, Src_size + (MP3_FRAMES_LAG * 4) + 8); Uint8 *rawbuf = (Uint8 *) malloc(rawbufsize + (MP3_FRAMES_LAG * 4) + 8); memset(rawbuf, 0, rawbufsize + (MP3_FRAMES_LAG * 4) + 8); ACMSTREAMHEADER Pack_Stream_Head; ZeroMemory(&Pack_Stream_Head, sizeof(ACMSTREAMHEADER)); Pack_Stream_Head.cbStruct = sizeof(ACMSTREAMHEADER); Pack_Stream_Head.pbSrc = (Uint8 *) Pack_Buf; Pack_Stream_Head.cbSrcLength = Src_size + (MP3_FRAMES_LAG * 2); Pack_Stream_Head.pbDst = rawbuf; Pack_Stream_Head.cbDstLength = rawbufsize; acmStreamPrepareHeader(Pack_Stream, &Pack_Stream_Head, 0); memcpy(Pack_Buf, Source, Src_size); acmStreamConvert(Pack_Stream, &Pack_Stream_Head, 0); Dest_Size = Pack_Stream_Head.cbDstLengthUsed; if(Dest_Size < Src_size) { memcpy(Dest, rawbuf, Dest_Size); } else { Dest_Size = 0; } acmStreamUnprepareHeader(Pack_Stream, &Pack_Stream_Head, 0); acmStreamClose(Pack_Stream, 0); if(rawbuf) free(rawbuf); if(Pack_Buf) free(Pack_Buf); return(Dest_Size); }
int ToAT3(short *Source, short *Dest, int Size, int BitRate) { int Src_size; int Dest_Size; int i; Wave_Format.wFormatTag = WAVE_FORMAT_PCM; Wave_Format.nChannels = 2; Wave_Format.cbSize = 0; Wave_Format.wBitsPerSample = 16; Wave_Format.nSamplesPerSec = 44100; Wave_Format.nBlockAlign = Wave_Format.nChannels * Wave_Format.wBitsPerSample / 8; Wave_Format.nAvgBytesPerSec = Wave_Format.nSamplesPerSec * Wave_Format.nBlockAlign; At3_Format.wfx.wFormatTag = 0x270; At3_Format.wfx.nChannels = 2; At3_Format.wfx.nSamplesPerSec = 44100; At3_Format.wfx.nAvgBytesPerSec = BitRate * 125; switch(BitRate) { case 66: At3_Format.wfx.nBlockAlign = 192; break; case 105: At3_Format.wfx.nBlockAlign = 304; break; case 132: At3_Format.wfx.nBlockAlign = 384; break; } At3_Format.wfx.wBitsPerSample = 0; At3_Format.wfx.cbSize = 0xe; At3_Format.wRevision = 1; At3_Format.nSamplesPerBlock = 0x800; At3_Format.abReserved[2] = 1; At3_Format.abReserved[4] = 1; At3_Format.abReserved[6] = 1; acmStreamOpen(&Pack_Stream, NULL, (LPWAVEFORMATEX) &Wave_Format, (LPWAVEFORMATEX) &At3_Format, NULL, 0, 0, 0); short *dwSource = (short *) malloc(Size * 2 + 8); memset(dwSource, 0, Size * 2 + 8); for(i = 0; i < Size / 2; i++) { dwSource[(i * 2)] = Source[i]; dwSource[(i * 2) + 1] = 0; } Source = dwSource; Size *= 2; Src_size = Size; unsigned long rawbufsize = 0; acmStreamSize(Pack_Stream, Src_size, &rawbufsize, ACM_STREAMSIZEF_SOURCE); Uint8 *Pack_Buf = (Uint8 *) malloc(Src_size + 8); memset(Pack_Buf, 0, Src_size + 8); Uint8 *rawbuf = (Uint8 *) malloc(rawbufsize + 8); memset(rawbuf, 0, rawbufsize + 8); ACMSTREAMHEADER Pack_Stream_Head; ZeroMemory(&Pack_Stream_Head, sizeof(ACMSTREAMHEADER)); Pack_Stream_Head.cbStruct = sizeof(ACMSTREAMHEADER); Pack_Stream_Head.pbSrc = (Uint8 *) Pack_Buf; Pack_Stream_Head.cbSrcLength = Src_size; Pack_Stream_Head.pbDst = rawbuf; Pack_Stream_Head.cbDstLength = rawbufsize; acmStreamPrepareHeader(Pack_Stream, &Pack_Stream_Head, 0); memcpy(Pack_Buf, Source, Src_size); acmStreamConvert(Pack_Stream, &Pack_Stream_Head, 0); Dest_Size = Pack_Stream_Head.cbDstLengthUsed; if(Dest_Size < Src_size) { memcpy(Dest, rawbuf, Dest_Size); } else { Dest_Size = 0; } acmStreamUnprepareHeader(Pack_Stream, &Pack_Stream_Head, 0); if(rawbuf) free(rawbuf); if(Pack_Buf) free(Pack_Buf); if(dwSource) free(dwSource); acmStreamClose(Pack_Stream, 0); return(Dest_Size); }
static HRESULT WINAPI ACMWrapper_Receive(TransformFilter *tf, IMediaSample *pSample) { ACMWrapperImpl* This = (ACMWrapperImpl*)tf; AM_MEDIA_TYPE amt; IMediaSample* pOutSample = NULL; DWORD cbDstStream, cbSrcStream; LPBYTE pbDstStream; LPBYTE pbSrcStream = NULL; ACMSTREAMHEADER ash; BOOL unprepare_header = FALSE, preroll; MMRESULT res; HRESULT hr; LONGLONG tStart = -1, tStop = -1, tMed; EnterCriticalSection(&This->tf.filter.csFilter); hr = IMediaSample_GetPointer(pSample, &pbSrcStream); if (FAILED(hr)) { ERR("Cannot get pointer to sample data (%x)\n", hr); LeaveCriticalSection(&This->tf.filter.csFilter); return hr; } preroll = (IMediaSample_IsPreroll(pSample) == S_OK); IMediaSample_GetTime(pSample, &tStart, &tStop); cbSrcStream = IMediaSample_GetActualDataLength(pSample); /* Prevent discontinuities when codecs 'absorb' data but not give anything back in return */ if (IMediaSample_IsDiscontinuity(pSample) == S_OK) { This->lasttime_real = tStart; This->lasttime_sent = tStart; } else if (This->lasttime_real == tStart) tStart = This->lasttime_sent; else WARN("Discontinuity\n"); tMed = tStart; TRACE("Sample data ptr = %p, size = %d\n", pbSrcStream, cbSrcStream); hr = IPin_ConnectionMediaType(This->tf.ppPins[0], &amt); if (FAILED(hr)) { ERR("Unable to retrieve media type\n"); LeaveCriticalSection(&This->tf.filter.csFilter); return hr; } ash.pbSrc = pbSrcStream; ash.cbSrcLength = cbSrcStream; while(hr == S_OK && ash.cbSrcLength) { hr = BaseOutputPinImpl_GetDeliveryBuffer((BaseOutputPin*)This->tf.ppPins[1], &pOutSample, NULL, NULL, 0); if (FAILED(hr)) { ERR("Unable to get delivery buffer (%x)\n", hr); LeaveCriticalSection(&This->tf.filter.csFilter); return hr; } IMediaSample_SetPreroll(pOutSample, preroll); hr = IMediaSample_SetActualDataLength(pOutSample, 0); assert(hr == S_OK); hr = IMediaSample_GetPointer(pOutSample, &pbDstStream); if (FAILED(hr)) { ERR("Unable to get pointer to buffer (%x)\n", hr); goto error; } cbDstStream = IMediaSample_GetSize(pOutSample); ash.cbStruct = sizeof(ash); ash.fdwStatus = 0; ash.dwUser = 0; ash.pbDst = pbDstStream; ash.cbDstLength = cbDstStream; if ((res = acmStreamPrepareHeader(This->has, &ash, 0))) { ERR("Cannot prepare header %d\n", res); goto error; } unprepare_header = TRUE; if (IMediaSample_IsDiscontinuity(pSample) == S_OK) { res = acmStreamConvert(This->has, &ash, ACM_STREAMCONVERTF_START); IMediaSample_SetDiscontinuity(pOutSample, TRUE); /* One sample could be converted to multiple packets */ IMediaSample_SetDiscontinuity(pSample, FALSE); } else { res = acmStreamConvert(This->has, &ash, 0); IMediaSample_SetDiscontinuity(pOutSample, FALSE); } if (res) { if(res != MMSYSERR_MOREDATA) ERR("Cannot convert data header %d\n", res); goto error; } TRACE("used in %u/%u, used out %u/%u\n", ash.cbSrcLengthUsed, ash.cbSrcLength, ash.cbDstLengthUsed, ash.cbDstLength); hr = IMediaSample_SetActualDataLength(pOutSample, ash.cbDstLengthUsed); assert(hr == S_OK); /* Bug in acm codecs? It apparantly uses the input, but doesn't necessarily output immediately kl*/ if (!ash.cbSrcLengthUsed) { WARN("Sample was skipped? Outputted: %u\n", ash.cbDstLengthUsed); ash.cbSrcLength = 0; goto error; } TRACE("Sample start time: %u.%03u\n", (DWORD)(tStart/10000000), (DWORD)((tStart/10000)%1000)); if (ash.cbSrcLengthUsed == cbSrcStream) { IMediaSample_SetTime(pOutSample, &tStart, &tStop); tStart = tMed = tStop; } else if (tStop != tStart) { tMed = tStop - tStart; tMed = tStart + tMed * ash.cbSrcLengthUsed / cbSrcStream; IMediaSample_SetTime(pOutSample, &tStart, &tMed); tStart = tMed; } else { ERR("No valid timestamp found\n"); IMediaSample_SetTime(pOutSample, NULL, NULL); } TRACE("Sample stop time: %u.%03u\n", (DWORD)(tStart/10000000), (DWORD)((tStart/10000)%1000)); LeaveCriticalSection(&This->tf.filter.csFilter); hr = BaseOutputPinImpl_Deliver((BaseOutputPin*)This->tf.ppPins[1], pOutSample); EnterCriticalSection(&This->tf.filter.csFilter); if (hr != S_OK && hr != VFW_E_NOT_CONNECTED) { if (FAILED(hr)) ERR("Error sending sample (%x)\n", hr); goto error; } error: if (unprepare_header && (res = acmStreamUnprepareHeader(This->has, &ash, 0))) ERR("Cannot unprepare header %d\n", res); unprepare_header = FALSE; ash.pbSrc += ash.cbSrcLengthUsed; ash.cbSrcLength -= ash.cbSrcLengthUsed; if (pOutSample) IMediaSample_Release(pOutSample); pOutSample = NULL; } This->lasttime_real = tStop; This->lasttime_sent = tMed; LeaveCriticalSection(&This->tf.filter.csFilter); return hr; }
static HRESULT WINAPI ACMStream_fnWrite(IAVIStream *iface, LONG start, LONG samples, LPVOID buffer, LONG buffersize, DWORD flags, LPLONG sampwritten, LPLONG byteswritten) { IAVIStreamImpl *This = (IAVIStreamImpl *)iface; HRESULT hr; ULONG size; TRACE("(%p,%d,%d,%p,%d,0x%08X,%p,%p)\n", iface, start, samples, buffer, buffersize, flags, sampwritten, byteswritten); /* clear return parameters if given */ if (sampwritten != NULL) *sampwritten = 0; if (byteswritten != NULL) *byteswritten = 0; /* check parameters */ if (buffer == NULL && (buffersize > 0 || samples > 0)) return AVIERR_BADPARAM; /* Have we write capability? */ if ((This->sInfo.dwCaps & AVIFILECAPS_CANWRITE) == 0) return AVIERR_READONLY; /* also need a compressor */ if (This->has == NULL) return AVIERR_NOCOMPRESSOR; /* map our sizes to pStream sizes */ size = buffersize; CONVERT_THIS_to_STREAM(&size); CONVERT_THIS_to_STREAM(&start); /* no bytes to write? -- short circuit */ if (size == 0) { return IAVIStream_Write(This->pStream, -1, samples, buffer, size, flags, sampwritten, byteswritten); } /* Need to free source buffer used for reading? */ if (This->acmStreamHdr.pbSrc != NULL) { HeapFree(GetProcessHeap(), 0, This->acmStreamHdr.pbSrc); This->acmStreamHdr.pbSrc = NULL; This->acmStreamHdr.dwSrcUser = 0; } /* Need bigger destination buffer? */ if (This->acmStreamHdr.pbDst == NULL || This->acmStreamHdr.dwDstUser < size) { if (This->acmStreamHdr.pbDst == NULL) This->acmStreamHdr.pbDst = HeapAlloc(GetProcessHeap(), 0, size); else This->acmStreamHdr.pbDst = HeapReAlloc(GetProcessHeap(), 0, This->acmStreamHdr.pbDst, size); if (This->acmStreamHdr.pbDst == NULL) return AVIERR_MEMORY; This->acmStreamHdr.dwDstUser = size; } This->acmStreamHdr.cbStruct = sizeof(This->acmStreamHdr); This->acmStreamHdr.cbSrcLengthUsed = 0; This->acmStreamHdr.cbDstLengthUsed = 0; This->acmStreamHdr.cbDstLength = This->acmStreamHdr.dwDstUser; /* need to prepare stream? */ This->acmStreamHdr.pbSrc = buffer; This->acmStreamHdr.cbSrcLength = buffersize; if ((This->acmStreamHdr.fdwStatus & ACMSTREAMHEADER_STATUSF_PREPARED) == 0) { if (acmStreamPrepareHeader(This->has, &This->acmStreamHdr, 0) != S_OK) { This->acmStreamHdr.pbSrc = NULL; This->acmStreamHdr.cbSrcLength = 0; return AVIERR_COMPRESSOR; } } /* now do the conversion */ /* FIXME: use ACM_CONVERTF_* flags */ if (acmStreamConvert(This->has, &This->acmStreamHdr, 0) != S_OK) hr = AVIERR_COMPRESSOR; else hr = AVIERR_OK; This->acmStreamHdr.pbSrc = NULL; This->acmStreamHdr.cbSrcLength = 0; if (FAILED(hr)) return hr; return IAVIStream_Write(This->pStream,-1,This->acmStreamHdr.cbDstLengthUsed / This->lpOutFormat->nBlockAlign,This->acmStreamHdr.pbDst, This->acmStreamHdr.cbDstLengthUsed,flags,sampwritten, byteswritten); }
BOOL ClSoundDS::CreateStreamBuffer(int handle,int arcFileNum,int playNum) { int size,stream,dataTopOffset,nextStream = -1; HRESULT hr; RiffHead riffHead; FmtHead fmtHead; PCMWAVEFORMAT pcmFormat; BOOL bOGG = FALSE; ReleaseSoundBuffer(handle); stream = lpReadFile->StreamOpenFileNum(arcFileNum,playNum,size); if(pack_bgmfile!=arcFileNum && pack_voice!=arcFileNum){ lpReadFile->StreamReadFile(arcFileNum,stream,(char *)&riffHead,sizeof(RiffHead)); if(strncmp(riffHead.riff,"RIFF",4) || strncmp(riffHead.type,"WAVE",4)){ lpReadFile->StreamCloseFile(arcFileNum,stream); return FALSE; } lpReadFile->StreamReadFile(arcFileNum,stream,(char *)&fmtHead,sizeof(FmtHead)); lpReadFile->StreamReadFile(arcFileNum,stream,(char *)&pcmFormat,sizeof(PCMWAVEFORMAT)); }else{ char chkFmt[4]; bOGG = TRUE; char *fname = lpReadFile->GetFileName(arcFileNum,playNum); lpReadFile->StreamReadFile(arcFileNum,stream,(char *)&chkFmt,4); lpReadFile->StreamSeekFile(arcFileNum,stream,0,FILE_BEGIN); if(0!=strncmp(chkFmt,"OggS",4))return FALSE; dataTopOffset = 0; if(NULL==lpDSBufferTop){ lpDSBufferTop = lpDSBufferTail = new ClSoundBuffer; }else{ lpDSBufferTail->lpNext = new ClSoundBuffer; lpDSBufferTail->lpNext->lpPrev = lpDSBufferTail; lpDSBufferTail = lpDSBufferTail->lpNext; } lpDSBufferTail->dataTopOffset = dataTopOffset; lpDSBufferTail->o_dat = new OggDec; if(_strnicmp(&fname[strlen(fname)-6],"_A.",3) == 0){ char workBuf[32]; strcpy(workBuf,fname); workBuf[strlen(fname)-5] = 'B'; nextStream = lpReadFile->SearchFile(arcFileNum,workBuf); } WAVEFORMATEX dstWF; char workBuf[4096]; lpReadFile->StreamReadFile(arcFileNum,stream,workBuf,4096); lpDSBufferTail->o_dat->GetWaveformat(&dstWF,workBuf); DSBUFFERDESC dsbdesc; ZeroMemory(&dsbdesc, sizeof(DSBUFFERDESC)); dsbdesc.dwSize = sizeof(DSBUFFERDESC); dsbdesc.dwFlags = DSBCAPS_STATIC | DSBCAPS_CTRLVOLUME | DSBCAPS_CTRLPOSITIONNOTIFY | DSBCAPS_LOCSOFTWARE | DSBCAPS_GLOBALFOCUS; dsbdesc.dwBufferBytes = dstWF.nAvgBytesPerSec *2; dsbdesc.lpwfxFormat = &dstWF; hr = lpDSound->CreateSoundBuffer(&dsbdesc, &lpTmpBuffer, NULL); lpTmpBuffer->QueryInterface(IID_IDirectSoundBuffer8,(LPVOID *)&lpDSBufferTail->lpDSBuffer); RELEASE(lpTmpBuffer); lpDSBufferTail->streamBlockSize = dstWF.nAvgBytesPerSec; } if(bOGG){ pcmFormat.wf.wFormatTag = WAVE_FORMAT_OGG; }else if(pcmFormat.wf.wFormatTag != WAVE_FORMAT_PCM){ WAVEFORMATEX dstWF; WAVEFORMATEX *pwfxInfo; WORD cbExtraAlloc; MMRESULT mmResult; lpReadFile->StreamReadFile(arcFileNum,stream,(char *)&cbExtraAlloc,sizeof(WORD)); pwfxInfo = (WAVEFORMATEX *)cl_malloc(sizeof(WAVEFORMATEX) +cbExtraAlloc); CopyMemory(pwfxInfo,&pcmFormat,sizeof(PCMWAVEFORMAT)); pwfxInfo->cbSize = cbExtraAlloc; lpReadFile->StreamReadFile(arcFileNum,stream,(char *)pwfxInfo +sizeof(WAVEFORMATEX),cbExtraAlloc); lpReadFile->StreamSeekFile(arcFileNum,stream,sizeof(RiffHead),FILE_BEGIN); lpReadFile->StreamReadFile(arcFileNum,stream,(char *)&fmtHead,sizeof(FmtHead)); dataTopOffset = sizeof(RiffHead) +sizeof(FmtHead); while(0!=strncmp(fmtHead.fmt,"data",4)){ lpReadFile->StreamSeekFile(arcFileNum,stream,fmtHead.size,FILE_CURRENT); dataTopOffset += fmtHead.size; lpReadFile->StreamReadFile(arcFileNum,stream,(char *)&fmtHead,sizeof(FmtHead)); dataTopOffset += sizeof(FmtHead); } ZeroMemory(&dstWF,sizeof(dstWF)); dstWF.wFormatTag = WAVE_FORMAT_PCM; mmResult = acmFormatSuggest( NULL, pwfxInfo, &dstWF, sizeof(dstWF), ACM_FORMATSUGGESTF_WFORMATTAG); if(mmResult != 0){ myOutputDebugString("オーディオ圧縮が無効です\n"); return FALSE; } if(NULL==lpDSBufferTop){ lpDSBufferTop = lpDSBufferTail = new ClSoundBuffer; }else{ lpDSBufferTail->lpNext = new ClSoundBuffer; lpDSBufferTail->lpNext->lpPrev = lpDSBufferTail; lpDSBufferTail = lpDSBufferTail->lpNext; } lpDSBufferTail->dataTopOffset = dataTopOffset; acmStreamOpen(&lpDSBufferTail->hAcm, NULL,pwfxInfo,&dstWF, NULL, 0L, 0L, ACM_STREAMOPENF_NONREALTIME); size = pwfxInfo->nAvgBytesPerSec; if(size % pwfxInfo->nBlockAlign){ size = (size/pwfxInfo->nBlockAlign +1)*pwfxInfo->nBlockAlign; } acmStreamSize(lpDSBufferTail->hAcm,size, &lpDSBufferTail->acmDst.dwStreamSize, ACM_STREAMSIZEF_SOURCE); acmStreamSize(lpDSBufferTail->hAcm, lpDSBufferTail->acmDst.dwStreamSize, &lpDSBufferTail->acmSrc.dwStreamSize, ACM_STREAMSIZEF_DESTINATION); acmStreamSize(lpDSBufferTail->hAcm, lpDSBufferTail->acmSrc.dwStreamSize, &lpDSBufferTail->acmDst.dwStreamSize, ACM_STREAMSIZEF_SOURCE); cl_free(pwfxInfo); if(fmtHead.size < lpDSBufferTail->acmSrc.dwStreamSize*2){ acmStreamClose(lpDSBufferTail->hAcm, 0); lpDSBufferTail->hAcm = NULL; lpReadFile->StreamCloseFile(arcFileNum,stream); if(lpDSBufferTop == lpDSBufferTail){ delete lpDSBufferTail; lpDSBufferTop = lpDSBufferTail = NULL; }else{ lpDSBufferTail = lpDSBufferTail->lpPrev; delete(lpDSBufferTail->lpNext); lpDSBufferTail->lpNext = NULL; } return -1; } lpDSBufferTail->acmSrc.lpStream = (LPBYTE)cl_malloc(lpDSBufferTail->acmSrc.dwStreamSize); lpDSBufferTail->acmDst.lpStream = (LPBYTE)cl_malloc(lpDSBufferTail->acmDst.dwStreamSize); ZeroMemory(&lpDSBufferTail->ash,sizeof(lpDSBufferTail->ash)); lpDSBufferTail->ash.cbStruct = sizeof(lpDSBufferTail->ash); lpDSBufferTail->ash.pbSrc = lpDSBufferTail->acmSrc.lpStream; lpDSBufferTail->ash.cbSrcLength = lpDSBufferTail->acmSrc.dwStreamSize; lpDSBufferTail->ash.dwSrcUser = lpDSBufferTail->acmSrc.dwStreamSize; lpDSBufferTail->ash.pbDst = lpDSBufferTail->acmDst.lpStream; lpDSBufferTail->ash.cbDstLength = lpDSBufferTail->acmDst.dwStreamSize; lpDSBufferTail->ash.dwDstUser = lpDSBufferTail->acmDst.dwStreamSize; acmStreamPrepareHeader(lpDSBufferTail->hAcm,&lpDSBufferTail->ash,0); DSBUFFERDESC dsbdesc; ZeroMemory(&dsbdesc, sizeof(DSBUFFERDESC)); dsbdesc.dwSize = sizeof(DSBUFFERDESC); dsbdesc.dwFlags = DSBCAPS_STATIC | DSBCAPS_CTRLVOLUME | DSBCAPS_CTRLPOSITIONNOTIFY | DSBCAPS_LOCSOFTWARE | DSBCAPS_GLOBALFOCUS; dsbdesc.dwBufferBytes = lpDSBufferTail->acmDst.dwStreamSize *2; dsbdesc.lpwfxFormat = &dstWF; hr = lpDSound->CreateSoundBuffer(&dsbdesc, &lpTmpBuffer, NULL); lpTmpBuffer->QueryInterface(IID_IDirectSoundBuffer8,(LPVOID *)&lpDSBufferTail->lpDSBuffer); RELEASE(lpTmpBuffer); lpDSBufferTail->streamBlockSize = lpDSBufferTail->acmDst.dwStreamSize; }else{ dataTopOffset = sizeof(RiffHead) +sizeof(FmtHead) +fmtHead.size; lpReadFile->StreamSeekFile(arcFileNum,stream,fmtHead.size -sizeof(PCMWAVEFORMAT),FILE_CURRENT); lpReadFile->StreamReadFile(arcFileNum,stream,(char *)&fmtHead,sizeof(FmtHead)); dataTopOffset += sizeof(FmtHead); while(0!=strncmp(fmtHead.fmt,"data",4)){ lpReadFile->StreamSeekFile(arcFileNum,stream,fmtHead.size,FILE_CURRENT); dataTopOffset += fmtHead.size; lpReadFile->StreamReadFile(arcFileNum,stream,(char *)&fmtHead,sizeof(FmtHead)); dataTopOffset += sizeof(FmtHead); } if(fmtHead.size < pcmFormat.wf.nAvgBytesPerSec*2){ lpReadFile->StreamCloseFile(arcFileNum,stream); return -1; } if(NULL==lpDSBufferTop){ lpDSBufferTop = lpDSBufferTail = new ClSoundBuffer; }else{ lpDSBufferTail->lpNext = new ClSoundBuffer; lpDSBufferTail->lpNext->lpPrev = lpDSBufferTail; lpDSBufferTail = lpDSBufferTail->lpNext; } lpDSBufferTail->dataTopOffset = dataTopOffset; DSBUFFERDESC dsbdesc; WAVEFORMATEX audioFmt; ZeroMemory(&dsbdesc, sizeof(DSBUFFERDESC)); dsbdesc.dwSize = sizeof(DSBUFFERDESC); dsbdesc.dwFlags = DSBCAPS_STATIC | DSBCAPS_CTRLVOLUME | DSBCAPS_CTRLPOSITIONNOTIFY | DSBCAPS_LOCSOFTWARE | DSBCAPS_GLOBALFOCUS; CopyMemory(&audioFmt,&pcmFormat,sizeof(WAVEFORMAT)); audioFmt.wBitsPerSample = pcmFormat.wBitsPerSample; audioFmt.cbSize = 0; dsbdesc.dwBufferBytes = audioFmt.nAvgBytesPerSec *2; dsbdesc.lpwfxFormat = &audioFmt; hr = lpDSound->CreateSoundBuffer(&dsbdesc, &lpTmpBuffer, NULL); lpTmpBuffer->QueryInterface(IID_IDirectSoundBuffer8,(LPVOID *)&lpDSBufferTail->lpDSBuffer); RELEASE(lpTmpBuffer); lpDSBufferTail->streamBlockSize = audioFmt.nAvgBytesPerSec; } lpDSBufferTail->arcFileNum = arcFileNum; lpDSBufferTail->musicNum = lpDSBufferTail->orgMusicNum = playNum; lpDSBufferTail->streamNum = stream; lpDSBufferTail->nextStream = nextStream; lpDSBufferTail->handle = handle; lpDSBufferTail->bufType = stream_sound; lpDSBufferTail->readFile = lpReadFile; lpDSBufferTail->wFormatTag = pcmFormat.wf.wFormatTag; return TRUE; } // ClSoundDS::CreateStreamBuffer
static int decode_audio(sh_audio_t *sh_audio,unsigned char *buf,int minlen,int maxlen) { ACMSTREAMHEADER ash; HRESULT hr; DWORD srcsize=0; DWORD len=minlen; acm_context_t *priv = sh_audio->context; acmStreamSize(priv->handle, len, &srcsize, ACM_STREAMSIZEF_DESTINATION); mp_msg(MSGT_WIN32,MSGL_DBG3,"acm says: srcsize=%ld (buffsize=%d) out_size=%ld\n",srcsize,sh_audio->a_in_buffer_size,len); if(srcsize<sh_audio->wf->nBlockAlign){ srcsize=sh_audio->wf->nBlockAlign; acmStreamSize(priv->handle, srcsize, &len, ACM_STREAMSIZEF_SOURCE); if(len>maxlen) len=maxlen; } // if(srcsize==0) srcsize=((WAVEFORMATEX *)&sh_audio->o_wf_ext)->nBlockAlign; if(srcsize>sh_audio->a_in_buffer_size) srcsize=sh_audio->a_in_buffer_size; // !!!!!! if(sh_audio->a_in_buffer_len<srcsize){ sh_audio->a_in_buffer_len+= demux_read_data(sh_audio->ds,&sh_audio->a_in_buffer[sh_audio->a_in_buffer_len], srcsize-sh_audio->a_in_buffer_len); } mp_msg(MSGT_WIN32,MSGL_DBG3,"acm convert %d -> %ld bytes\n",sh_audio->a_in_buffer_len,len); memset(&ash, 0, sizeof(ash)); ash.cbStruct=sizeof(ash); ash.fdwStatus=0; ash.dwUser=0; ash.pbSrc=sh_audio->a_in_buffer; ash.cbSrcLength=sh_audio->a_in_buffer_len; ash.pbDst=buf; ash.cbDstLength=len; hr=acmStreamPrepareHeader(priv->handle,&ash,0); if(hr){ mp_msg(MSGT_WIN32,MSGL_V,"ACM_Decoder: acmStreamPrepareHeader error %d\n",(int)hr); return -1; } hr=acmStreamConvert(priv->handle,&ash,0); if(hr){ mp_msg(MSGT_WIN32,MSGL_DBG2,"ACM_Decoder: acmStreamConvert error %d\n",(int)hr); switch(hr) { case ACMERR_NOTPOSSIBLE: case ACMERR_UNPREPARED: mp_msg(MSGT_WIN32, MSGL_DBG2, "ACM_Decoder: acmStreamConvert error: probarly not initialized!\n"); } // return -1; } mp_msg(MSGT_WIN32,MSGL_DBG2,"acm converted %ld -> %ld\n",ash.cbSrcLengthUsed,ash.cbDstLengthUsed); if(ash.cbSrcLengthUsed>=sh_audio->a_in_buffer_len){ sh_audio->a_in_buffer_len=0; } else { sh_audio->a_in_buffer_len-=ash.cbSrcLengthUsed; memcpy(sh_audio->a_in_buffer,&sh_audio->a_in_buffer[ash.cbSrcLengthUsed],sh_audio->a_in_buffer_len); } len=ash.cbDstLengthUsed; hr=acmStreamUnprepareHeader(priv->handle,&ash,0); if(hr){ mp_msg(MSGT_WIN32,MSGL_V,"ACM_Decoder: acmStreamUnprepareHeader error %d\n",(int)hr); } return len; }
// // FillBuffer // // Stuffs the buffer with data HRESULT CSynthStream::FillBuffer(IMediaSample *pms) { CheckPointer(pms,E_POINTER); BYTE *pData; HRESULT hr = pms->GetPointer(&pData); if (FAILED(hr)) { return hr; } // This function must hold the state lock because it calls // FillPCMAudioBuffer(). CAutoLock lStateLock(m_pParent->pStateLock()); // This lock must be held because this function uses // m_dwTempPCMBufferSize, m_hPCMToMSADPCMConversionStream, // m_rtSampleTime, m_fFirstSampleDelivered and // m_llSampleMediaTimeStart. CAutoLock lShared(&m_cSharedState); WAVEFORMATEX* pwfexCurrent = (WAVEFORMATEX*)m_mt.Format(); if (WAVE_FORMAT_PCM == pwfexCurrent->wFormatTag) { m_Synth->FillPCMAudioBuffer(*pwfexCurrent, pData, pms->GetSize()); hr = pms->SetActualDataLength(pms->GetSize()); if (FAILED(hr)) return hr; } else { // This filter only supports two audio formats: PCM and ADPCM. ASSERT(WAVE_FORMAT_ADPCM == pwfexCurrent->wFormatTag); // Create PCM audio samples and then compress them. We use the // Audio Compression Manager (ACM) API to convert the samples to // the ADPCM format. ACMSTREAMHEADER ACMStreamHeader; ACMStreamHeader.cbStruct = sizeof(ACMStreamHeader); ACMStreamHeader.fdwStatus = 0; ACMStreamHeader.dwUser = 0; ACMStreamHeader.cbSrcLength = m_dwTempPCMBufferSize; ACMStreamHeader.cbSrcLengthUsed = 0; ACMStreamHeader.dwSrcUser = 0; ACMStreamHeader.pbDst = pData; ACMStreamHeader.cbDstLength = pms->GetSize(); ACMStreamHeader.cbDstLengthUsed = 0; ACMStreamHeader.dwDstUser = 0; ACMStreamHeader.pbSrc = new BYTE[m_dwTempPCMBufferSize]; if (NULL == ACMStreamHeader.pbSrc) { return E_OUTOFMEMORY; } WAVEFORMATEX wfexPCMAudio; DerivePCMFormatFromADPCMFormatStructure(*pwfexCurrent, &wfexPCMAudio); m_Synth->FillPCMAudioBuffer(wfexPCMAudio, ACMStreamHeader.pbSrc, ACMStreamHeader.cbSrcLength); MMRESULT mmr = acmStreamPrepareHeader(m_hPCMToMSADPCMConversionStream, &ACMStreamHeader, 0); // acmStreamPrepareHeader() returns 0 if no errors occur. if (mmr != 0) { delete [] ACMStreamHeader.pbSrc; return E_FAIL; } mmr = acmStreamConvert(m_hPCMToMSADPCMConversionStream, &ACMStreamHeader, ACM_STREAMCONVERTF_BLOCKALIGN); MMRESULT mmrUnprepare = acmStreamUnprepareHeader(m_hPCMToMSADPCMConversionStream, &ACMStreamHeader, 0); delete [] ACMStreamHeader.pbSrc; // acmStreamConvert() andacmStreamUnprepareHeader() returns 0 if no errors occur. if ((mmr != 0) || (mmrUnprepare != 0)) { return E_FAIL; } hr = pms->SetActualDataLength(ACMStreamHeader.cbDstLengthUsed); if (FAILED(hr)) { return hr; } } // Set the sample's time stamps. CRefTime rtStart = m_rtSampleTime; m_rtSampleTime = rtStart + (REFERENCE_TIME)(UNITS * pms->GetActualDataLength()) / (REFERENCE_TIME)pwfexCurrent->nAvgBytesPerSec; hr = pms->SetTime((REFERENCE_TIME*)&rtStart, (REFERENCE_TIME*)&m_rtSampleTime); if (FAILED(hr)) { return hr; } // Set the sample's properties. hr = pms->SetPreroll(FALSE); if (FAILED(hr)) { return hr; } hr = pms->SetMediaType(NULL); if (FAILED(hr)) { return hr; } hr = pms->SetDiscontinuity(!m_fFirstSampleDelivered); if (FAILED(hr)) { return hr; } hr = pms->SetSyncPoint(!m_fFirstSampleDelivered); if (FAILED(hr)) { return hr; } LONGLONG llMediaTimeStart = m_llSampleMediaTimeStart; DWORD dwNumAudioSamplesInPacket = (pms->GetActualDataLength() * BITS_PER_BYTE) / (pwfexCurrent->nChannels * pwfexCurrent->wBitsPerSample); LONGLONG llMediaTimeStop = m_llSampleMediaTimeStart + dwNumAudioSamplesInPacket; hr = pms->SetMediaTime(&llMediaTimeStart, &llMediaTimeStop); if (FAILED(hr)) { return hr; } m_llSampleMediaTimeStart = llMediaTimeStop; m_fFirstSampleDelivered = TRUE; return NOERROR; }
AudioConverterStream::AudioConverterStream(WaveFormat sourceFormat, WaveFormat destFormat, WaveFormat** prevSourceFormats , int numPrevSourceFormats) : outBuffer(nullptr), outSize(0), outBufferAllocated(0), stream(nullptr), isProxy(false), m_failed(false), subConverter(nullptr), startOffset(0) { if (sourceFormat.GetSize() == destFormat.GetSize()) { LPWAVEFORMATEX src = static_cast<LPWAVEFORMATEX>(sourceFormat); LPWAVEFORMATEX dest = static_cast<LPWAVEFORMATEX>(destFormat); if (memcmp(src, dest, sourceFormat.GetSize()) == 0) { // Source and destination formats are identical, so we'll just pass the input through to // the output isProxy = true; return; } } MMRESULT mm = acmStreamOpen(&stream, nullptr, sourceFormat, destFormat, nullptr, 0, 0, ACM_STREAMOPENF_NONREALTIME); if (mm != MMSYSERR_NOERROR) { // Not supported directly, so try letting ACM suggest an intermediate format, so we can // perform the conversion in multiple stages. WaveFormat intermediateFormat; // This weird looping structure is because we must try all combinations of four flags, and // even when a given combination yields a valid suggestion, we must be able to backtrack and // continue looping (tryMoreSuggestions) if the conversion based on that suggestion later // fails or dead-ends. bool foundSuggest = false; int chan, samp, bits, form, done; for (done = 0; done < 1 && !foundSuggest; foundSuggest ? 0 : ++done) { for (chan = 1; chan >= 0 && !foundSuggest; foundSuggest ? 0 : --chan) { for (samp = 1; samp >= 0 && !foundSuggest; foundSuggest ? 0 : --samp) { for (bits = 0; bits <= 1 && !foundSuggest; foundSuggest ? 0 : ++bits) { for (form = 0; form <= 1 && !foundSuggest; foundSuggest ? 0 : ++form) { int flags = 0; flags |= chan ? ACM_FORMATSUGGESTF_NCHANNELS : 0; flags |= samp ? ACM_FORMATSUGGESTF_NSAMPLESPERSEC : 0; flags |= bits ? ACM_FORMATSUGGESTF_WBITSPERSAMPLE : 0; flags |= form ? ACM_FORMATSUGGESTF_WFORMATTAG : 0; intermediateFormat = destFormat; MMRESULT mmSuggest = acmFormatSuggest(NULL, sourceFormat, intermediateFormat, intermediateFormat.GetSize(), flags); if (mmSuggest == MMSYSERR_NOERROR) { // Got a possibly-valid suggestion, but it might be a suggestion to // do absolutely nothing (which would be bad), so we first make sure // there's some sort of change involved: if (!FormatsMatch(sourceFormat, intermediateFormat)) { // We got a suggestion foundSuggest = true; // Now check to see if it's identical to a previous conversion // state. If it is, then we'll revert foundSuggest to false to // prevent endless conversion cycles. for (int prev = 0; prev < numPrevSourceFormats && prevSourceFormats && foundSuggest; prev++) { WaveFormat& oldFormat = *prevSourceFormats[prev]; if (FormatsMatch(oldFormat, intermediateFormat)) { // We already went through this exact format foundSuggest = false; } } } } tryMoreSuggestions: continue; } } } } } if (!foundSuggest) { m_failed = true; return; } // we'll handle conversion to the intermediate format mm = acmStreamOpen(&stream, nullptr, sourceFormat, intermediateFormat, nullptr, 0, 0, ACM_STREAMOPENF_NONREALTIME); if (mm != MMSYSERR_NOERROR) { if (!done) { foundSuggest = false; goto tryMoreSuggestions; // continue the search } // reached dead end m_failed = true; return; } // create temporary updated conversion history for cycle prevention size_t prevSize = sizeof(WaveFormat*) * (numPrevSourceFormats + 1); WaveFormat** prevFormats = static_cast<WaveFormat**>(alloca(prevSize)); if (prevSourceFormats) { memcpy(prevFormats, prevSourceFormats, prevSize); } prevFormats[numPrevSourceFormats] = &sourceFormat; // delegate the rest of the conversion to a new converter (recursive construction) subConverter = new AudioConverterStream(intermediateFormat, destFormat, prevFormats, numPrevSourceFormats + 1); if (subConverter->m_failed) { delete subConverter; subConverter = nullptr; if (!done) { foundSuggest = false; goto tryMoreSuggestions; // continue the search } // reached dead end m_failed = true; return; } } // prepare the stream header memset(&header, 0, sizeof(ACMSTREAMHEADER)); header.cbStruct = sizeof(ACMSTREAMHEADER); header.pbSrc = inWorkBuffer; header.cbSrcLength = sizeof(inWorkBuffer); header.pbDst = outWorkBuffer; header.cbDstLength = sizeof(outWorkBuffer); mm = acmStreamPrepareHeader(stream, &header, 0); if (mm != MMSYSERR_NOERROR) { m_failed = true; } }
void audio_resampler_acm::open( void ) { MMRESULT err; // // Opens ACM stream // err = acmStreamOpen( &acm_stream, 0, &wformat_src, &wformat_dst, 0, 0, 0, ACM_STREAMOPENF_NONREALTIME ); if ( err != MMSYSERR_NOERROR ) { //TODO: throw error MessageBox( 0, _T("acmOpen error: %i"), _T("ERROR"), MB_ICONERROR ); } // // Calcs source buffer length // src_buflen = ( unsigned int ) (( float )audfmt_in.byte_rate() * ( float )buf_secs ); // // Calcs destination source buffer length // with help of ACM apis // err = acmStreamSize( acm_stream, src_buflen, &dst_buflen, ACM_STREAMSIZEF_SOURCE ); if ( err != MMSYSERR_NOERROR ) { //TODO: throw error MessageBox( 0, _T("acmStreamSize error"), _T("ERROR"), MB_ICONERROR ); } // // Initialize ACMSTREAMHEADER structure, // and alloc memory for source and destination // buffers. // acm_header.fdwStatus = 0; acm_header.dwUser = 0; acm_header.pbSrc = ( LPBYTE ) new BYTE [ src_buflen ]; acm_header.cbSrcLength = src_buflen; acm_header.cbSrcLengthUsed = 0; acm_header.dwSrcUser = src_buflen; acm_header.pbDst = ( LPBYTE ) new BYTE [ dst_buflen ]; acm_header.cbDstLength = dst_buflen; acm_header.cbDstLengthUsed = 0; acm_header.dwDstUser = dst_buflen; // // Give ACMSTREAMHEADER initialized correctly to the // driver. // err = acmStreamPrepareHeader( acm_stream, &acm_header, 0L ); if ( err != MMSYSERR_NOERROR ) { //TODO: throw error MessageBox( 0, _T("acmStreamPrepareHeader error"), _T("ERROR"), MB_ICONERROR ); } // // ACM stream successfully opened. // stream_opened = true; }
bool setupAudio() { int ret; //read audio stream info; specifically, we need the encoded chunksize AVISTREAMINFO audioStreamInfo; AVIStreamInfo(audioStream,&audioStreamInfo,sizeof(AVISTREAMINFO)); audioChunkSize = audioStreamInfo.dwSuggestedBufferSize; audioSampleCount = audioStreamInfo.dwLength; audioEncoded = new char[audioChunkSize]; //read the audio streamformat info LONG formatSize; AVIStreamReadFormat(audioStream,AVIStreamStart(audioStream),0,&formatSize); char *format = (char *)malloc(formatSize); AVIStreamReadFormat(audioStream,AVIStreamStart(audioStream),format,&formatSize); WAVEFORMATEX *wfxEncoded = (WAVEFORMATEX *)format; //construct a descriptor for the format we desire to get out of the decoder //note that we have to use the same samplerate as the encoded format indicates //since acm can't change the samplerate in one fell swoop wfxDecoded.cbSize = sizeof(WAVEFORMATEX); wfxDecoded.nChannels = wfxEncoded->nChannels; wfxDecoded.wFormatTag = WAVE_FORMAT_PCM; wfxDecoded.nSamplesPerSec = wfxEncoded->nSamplesPerSec; wfxDecoded.wBitsPerSample = 16; wfxDecoded.nBlockAlign = wfxDecoded.wBitsPerSample/8 * wfxEncoded->nChannels; wfxDecoded.nAvgBytesPerSec = wfxDecoded.nBlockAlign * wfxDecoded.nSamplesPerSec; sampleSize = wfxDecoded.nBlockAlign; //try to get a converter from the encoded data to the decoded data ret = acmStreamOpen(&acmStream,0,wfxEncoded,&wfxDecoded,0,0,0,0); //now we're done with wfxEncoded free(format); if(ret) { delete[] audioEncoded; AVIStreamClose(audioStream); return false; } //decide on a playback buffer size //make each buffer 1/2sec playBufferSamples = wfxDecoded.nSamplesPerSec / 2; playBufferSize = playBufferSamples * sampleSize; //hurry and try to create the output stream. //if we can't do that, then everything that follows is pointless. int mode = 0; if(wfxDecoded.wBitsPerSample == 8) mode |= FSOUND_8BITS; else if(wfxDecoded.wBitsPerSample == 16) mode |= FSOUND_16BITS; if(wfxDecoded.nChannels == 1) mode |= FSOUND_MONO; else mode |= FSOUND_STEREO; #ifdef SND_USE_FMOD fmod_stream = FSOUND_Stream_Create(win_movie_fmod_streamCallback,playBufferSize,mode,wfxDecoded.nSamplesPerSec,(int)this); if(!fmod_stream) { acmStreamClose(acmStream,0); delete[] audioEncoded; AVIStreamClose(audioStream); err("Error creating fmod stream for movieplayback. Please report this case so we can improve the robustness of the movie player!"); return false; } #endif //find out how large a decode buffer we need for the encode buffer chunksize acmStreamSize(acmStream,audioChunkSize,&decodeBufferSize,ACM_STREAMSIZEF_SOURCE); decodeBufferSamples = decodeBufferSize / sampleSize; //allocate the decode buffer audioDecoded = new char[decodeBufferSize]; //prep the decode operation audioStreamHeader.cbStruct = sizeof(ACMSTREAMHEADER); audioStreamHeader.fdwStatus = 0; audioStreamHeader.pbSrc = (LPBYTE)audioEncoded; audioStreamHeader.cbSrcLength = audioChunkSize; audioStreamHeader.pbDst = (LPBYTE)audioDecoded; audioStreamHeader.cbDstLength = decodeBufferSize; ret = acmStreamPrepareHeader(acmStream,&audioStreamHeader,0); if(ret) { delete[] audioDecoded; acmStreamClose(acmStream,0); delete[] audioEncoded; AVIStreamClose(audioStream); return false; } #ifdef SND_USE_FMOD //finally we're ready to start the audio stream FSOUND_Stream_Play(FSOUND_FREE,fmod_stream); #endif return true; }
void ConvertNode(HWND hwnd, AFile* node, const char* fname, WORD tag) { FSHandle* file; HANDLE wavfile; RIFFHeader riffhdr; ChunkHeader chunkhdr; DWORD riffsize,factsize,datasize,written,rate,buffpos,pcmsize,dstsize,sizeToFill,sizeFilled; DWORD pos_riffsize,pos_factsize,pos_datasize; WORD channels,bits; char str[MAX_PATH+100],*pcmBuffer=NULL,*dstBuffer=NULL; LPWAVEFORMATEX pwfex; ACMFORMATTAGDETAILS aftd={0}; MMRESULT mmr; WAVEFORMATEX wfexPCM; HACMSTREAM hACMStream; ACMSTREAMHEADER acmshdr; if ((file=FSOpenForPlayback(hwnd,node,&rate,&channels,&bits))==NULL) return; wfexPCM.wFormatTag=WAVE_FORMAT_PCM; wfexPCM.nChannels=channels; wfexPCM.nSamplesPerSec=rate; wfexPCM.wBitsPerSample=bits; wfexPCM.cbSize=0; wfexPCM.nBlockAlign=channels*(bits/8); wfexPCM.nAvgBytesPerSec=rate*wfexPCM.nBlockAlign; switch (tag) { case WAVE_FORMAT_PCM: pwfex=NULL; // ??? hACMStream=NULL; dstBuffer=NULL; // ??? pcmBuffer=(char*)GlobalAlloc(GPTR,BUFFERSIZE); break; default: aftd.cbStruct=sizeof(aftd); aftd.dwFormatTag=tag; mmr=acmFormatTagDetails(NULL,&aftd,ACM_FORMATTAGDETAILSF_LARGESTSIZE); if (mmr!=MMSYSERR_NOERROR) { AFPLUGIN(node)->ShutdownPlayback(file); FSCloseFile(file); wsprintf(str,"Failed to get details for wave format tag: 0x%X",tag); ReportMMError(hwnd,mmr,str); return; } pwfex=(LPWAVEFORMATEX)LocalAlloc(LPTR,aftd.cbFormatSize); pwfex->wFormatTag=tag; mmr=acmFormatSuggest(NULL,&wfexPCM,pwfex,aftd.cbFormatSize,ACM_FORMATSUGGESTF_WFORMATTAG); if (mmr!=MMSYSERR_NOERROR) { LocalFree(pwfex); AFPLUGIN(node)->ShutdownPlayback(file); FSCloseFile(file); wsprintf(str,"No format suggested for wave format tag: 0x%X",tag); ReportMMError(hwnd,mmr,str); return; } mmr=acmStreamOpen(&hACMStream,NULL,&wfexPCM,pwfex,NULL,0,0,ACM_STREAMOPENF_NONREALTIME); if (mmr!=MMSYSERR_NOERROR) { LocalFree(pwfex); AFPLUGIN(node)->ShutdownPlayback(file); FSCloseFile(file); wsprintf(str,"Failed to open conversion stream for wave format tag: 0x%X",tag); ReportMMError(hwnd,mmr,str); return; } if (acmStreamSize(hACMStream,BUFFERSIZE,&dstsize,ACM_STREAMSIZEF_SOURCE)!=MMSYSERR_NOERROR) dstsize=BUFFERSIZE; pcmBuffer=(char*)GlobalAlloc(GPTR,BUFFERSIZE); dstBuffer=(char*)GlobalAlloc(GPTR,dstsize); memset(&acmshdr,0x00,sizeof(acmshdr)); // ??? acmshdr.cbStruct=sizeof(ACMSTREAMHEADER); acmshdr.fdwStatus=0; acmshdr.pbSrc=pcmBuffer; acmshdr.cbSrcLength=BUFFERSIZE; acmshdr.cbSrcLengthUsed=0; acmshdr.pbDst=dstBuffer; acmshdr.cbDstLength=dstsize; acmshdr.cbDstLengthUsed=0; mmr=acmStreamPrepareHeader(hACMStream,&acmshdr,0L); if (mmr!=MMSYSERR_NOERROR) { GlobalFree(dstBuffer); GlobalFree(pcmBuffer); acmStreamClose(hACMStream,0); LocalFree(pwfex); AFPLUGIN(node)->ShutdownPlayback(file); FSCloseFile(file); ReportMMError(hwnd,mmr,"Failed to prepare conversion stream header."); return; } acmshdr.cbSrcLength=0; } if (!EnsureDirPresence(fname)) { if (hACMStream!=NULL) { acmStreamUnprepareHeader(hACMStream,&acmshdr,0L); acmStreamClose(hACMStream,0); } GlobalFree(dstBuffer); GlobalFree(pcmBuffer); LocalFree(pwfex); AFPLUGIN(node)->ShutdownPlayback(file); FSCloseFile(file); return; } wavfile=CreateFile(fname, GENERIC_WRITE, FILE_SHARE_READ, NULL, CREATE_ALWAYS, FILE_ATTRIBUTE_NORMAL, NULL ); if (wavfile==INVALID_HANDLE_VALUE) { if (hACMStream!=NULL) { acmStreamUnprepareHeader(hACMStream,&acmshdr,0L); acmStreamClose(hACMStream,0); } GlobalFree(dstBuffer); GlobalFree(pcmBuffer); LocalFree(pwfex); AFPLUGIN(node)->ShutdownPlayback(file); FSCloseFile(file); wsprintf(str,"Cannot open WAV file: %s",fname); ReportError(hwnd,str,NULL); return; } ShowProgressHeaderMsg(fname); datasize=0; factsize=0; ShowProgressStateMsg("Writing RIFF header..."); SetFilePointer(wavfile,0,NULL,FILE_BEGIN); lstrcpy(riffhdr.riffid,IDSTR_RIFF); lstrcpy(riffhdr.rifftype,IDSTR_WAVE); riffhdr.riffsize=0; WriteFile(wavfile,&riffhdr,sizeof(RIFFHeader),&written,NULL); pos_riffsize=SetFilePointer(wavfile,0,NULL,FILE_CURRENT)-sizeof(riffhdr.rifftype)-sizeof(riffhdr.riffsize); CorrectOddPos(wavfile); ShowProgressStateMsg("Writing fmt chunk..."); lstrcpy(chunkhdr.id,IDSTR_fmt); switch (tag) { case WAVE_FORMAT_PCM: chunkhdr.size=sizeof(wfexPCM); WriteFile(wavfile,&chunkhdr,sizeof(chunkhdr),&written,NULL); WriteFile(wavfile,&wfexPCM,chunkhdr.size,&written,NULL); CorrectOddPos(wavfile); break; default: chunkhdr.size=aftd.cbFormatSize; WriteFile(wavfile,&chunkhdr,sizeof(chunkhdr),&written,NULL); WriteFile(wavfile,pwfex,chunkhdr.size,&written,NULL); CorrectOddPos(wavfile); lstrcpy(chunkhdr.id,IDSTR_fact); chunkhdr.size=sizeof(factsize); WriteFile(wavfile,&chunkhdr,sizeof(chunkhdr),&written,NULL); pos_factsize=SetFilePointer(wavfile,0,NULL,FILE_CURRENT); WriteFile(wavfile,&factsize,sizeof(factsize),&written,NULL); CorrectOddPos(wavfile); } lstrcpy(chunkhdr.id,IDSTR_data); chunkhdr.size=datasize; WriteFile(wavfile,&chunkhdr,sizeof(chunkhdr),&written,NULL); pos_datasize=SetFilePointer(wavfile,0,NULL,FILE_CURRENT)-sizeof(datasize); while (1) { if (IsCancelled()) break; switch (tag) { case WAVE_FORMAT_PCM: sizeToFill=BUFFERSIZE; sizeFilled=0; break; default: if (acmshdr.cbSrcLengthUsed!=0L) { memmove(pcmBuffer,pcmBuffer+acmshdr.cbSrcLengthUsed,acmshdr.cbSrcLength-acmshdr.cbSrcLengthUsed); acmshdr.cbSrcLength-=acmshdr.cbSrcLengthUsed; } if (acmshdr.cbSrcLength<BUFFERSIZE) sizeToFill=BUFFERSIZE-acmshdr.cbSrcLength; sizeFilled=acmshdr.cbSrcLength; } wsprintf(str,"Converting %s data block to PCM...",file->node->afID); ShowProgressStateMsg(str); pcmsize=0; if (sizeToFill>0) pcmsize=AFPLUGIN(node)->FillPCMBuffer(file,pcmBuffer+sizeFilled,sizeToFill,&buffpos); if (tag==WAVE_FORMAT_PCM) { if (pcmsize==0L) break; ShowProgressStateMsg("Writing WAV data block..."); WriteFile(wavfile,pcmBuffer,pcmsize,&written,NULL); if (written!=pcmsize) { ReportError(hwnd,"Failure writing WAV file.",NULL); SetCancelFlag(); break; } datasize+=written; } else { acmshdr.cbSrcLength+=pcmsize; if (acmshdr.cbSrcLength==0L) break; acmshdr.fdwStatus^=ACMSTREAMHEADER_STATUSF_DONE; acmshdr.cbSrcLengthUsed=0; acmshdr.cbDstLength=dstsize; acmshdr.cbDstLengthUsed=0; wsprintf(str,"Compressing PCM data block..."); ShowProgressStateMsg(str); mmr=acmStreamConvert(hACMStream,&acmshdr,ACM_STREAMCONVERTF_BLOCKALIGN); if (mmr!=MMSYSERR_NOERROR) { CloseHandle(wavfile); DeleteFile(fname); acmStreamUnprepareHeader(hACMStream,&acmshdr,0L); acmStreamClose(hACMStream,0); GlobalFree(dstBuffer); GlobalFree(pcmBuffer); LocalFree(pwfex); AFPLUGIN(node)->ShutdownPlayback(file); FSCloseFile(file); ReportMMError(hwnd,mmr,"Error during compression."); return; } if (acmshdr.cbSrcLengthUsed==0L) { acmshdr.fdwStatus^=ACMSTREAMHEADER_STATUSF_DONE; acmStreamConvert(hACMStream,&acmshdr,0L); } factsize+=acmshdr.cbSrcLengthUsed/wfexPCM.nBlockAlign; ShowProgressStateMsg("Writing WAV data block..."); WriteFile(wavfile,dstBuffer,acmshdr.cbDstLengthUsed,&written,NULL); if (written!=acmshdr.cbDstLengthUsed) { ReportError(hwnd,"Failure writing WAV file.",NULL); SetCancelFlag(); break; } datasize+=written; } ShowProgress(FSGetFilePointer(file),FSGetFileSize(file)); // ??? } if (IsCancelled()) { ShowProgressStateMsg("Deleting WAV file..."); CloseHandle(wavfile); DeleteFile(fname); } else { CorrectOddPos(wavfile); ShowProgressStateMsg("Rewriting WAV header..."); riffsize=GetFileSize(wavfile,NULL)-8; SetFilePointer(wavfile,pos_riffsize,NULL,FILE_BEGIN); WriteFile(wavfile,&riffsize,sizeof(riffsize),&written,NULL); if (tag!=WAVE_FORMAT_PCM) { SetFilePointer(wavfile,pos_factsize,NULL,FILE_BEGIN); WriteFile(wavfile,&factsize,sizeof(factsize),&written,NULL); } SetFilePointer(wavfile,pos_datasize,NULL,FILE_BEGIN); WriteFile(wavfile,&datasize,sizeof(datasize),&written,NULL); CloseHandle(wavfile); } if (hACMStream!=NULL) { ShowProgressStateMsg("Closing conversion stream..."); acmStreamUnprepareHeader(hACMStream,&acmshdr,0L); acmStreamClose(hACMStream,0); } ShowProgressStateMsg("Freeing conversion buffers..."); GlobalFree(dstBuffer); GlobalFree(pcmBuffer); LocalFree(pwfex); wsprintf(str,"Shutting down %s decoder...",file->node->afID); ShowProgressStateMsg(str); AFPLUGIN(node)->ShutdownPlayback(file); FSCloseFile(file); }
DWORD CPlayer::WavPlayerThread() { BOOL fFlush = FALSE; if (m_pwfxSrc->wFormatTag == WAVE_FORMAT_PCM) { // PCM DWORD cbInBuf, cbRead; while (TRUE) { // 停止フラグのチェック if (m_fStop) return RET_STOP; if (!m_pOutHdr) m_pOutHdr = m_Output.GetBuffer(); { // Critical Sectionのセット CAutoLock lock(&m_csecThread); if (m_fSeek) { if (m_Status == MAP_STATUS_PLAY) m_fPlay = TRUE; m_Reverb.Reset(); m_Echo.Reset(); m_BassBoost.Reset(); m_3DChorus.Reset(); m_Output.Reset(); m_fSeek = FALSE; m_pOutHdr = NULL; continue; } // 読み込み cbRead = min(m_dwDataSize - m_dwCurrentSize, m_cbOutBuf); if (!m_Reader.Read((LPBYTE)m_pOutHdr->lpData, cbRead, &cbInBuf) || !cbInBuf) { if (GetLastError() != ERROR_SUCCESS) return RET_ERROR; fFlush = TRUE; } Preamp((LPBYTE)m_pOutHdr->lpData, cbInBuf); m_dwCurrentSize += cbInBuf; } if (fFlush) return RET_EOF; OutputBuffer(m_pOutHdr, cbInBuf); m_cbOutBufLeft = 0; m_pOutHdr = NULL; if (m_dwCurrentSize >= m_dwDataSize) return RET_EOF; if (m_fSuppress) return RET_EOF; } } else { // PCM 以外 // デコード開始 BYTE bRead[WAV_FILE_BUFF_LEN]; DWORD cbInBuf, cbRead; while (TRUE) { // 停止フラグのチェック if (m_fStop) return RET_STOP; // 出力バッファの確保 if (m_pOutHdr) { OutputBuffer(m_pOutHdr, m_cbOutBuf - m_cbOutBufLeft); m_cbOutBufLeft = 0; m_pOutHdr = NULL; } if (m_fSuppress) return RET_EOF; m_pOutHdr = m_Output.GetBuffer(); m_cbOutBufLeft = m_cbOutBuf; { // Critical Sectionのセット CAutoLock lock(&m_csecThread); if (m_fSeek) { if (m_Status == MAP_STATUS_PLAY) m_fPlay = TRUE; m_Reverb.Reset(); m_Echo.Reset(); m_BassBoost.Reset(); m_Output.Reset(); m_fSeek = FALSE; m_pOutHdr = NULL; acmStreamReset(m_hAcm, 0); continue; } acmStreamSize(m_hAcm, m_cbOutBufLeft, &cbRead, ACM_STREAMSIZEF_DESTINATION); // 読み込み cbRead = min(m_dwDataSize - m_dwCurrentSize, cbRead); cbRead = min(WAV_FILE_BUFF_LEN, cbRead); if (!m_Reader.Read(bRead, cbRead, &cbInBuf) || !cbInBuf) { if (GetLastError() != ERROR_SUCCESS) return RET_ERROR; fFlush = TRUE; } m_dwCurrentSize += cbInBuf; } if (fFlush) return RET_EOF; // ヘッダの確保 ACMSTREAMHEADER ash; memset(&ash, 0, sizeof(ash)); ash.cbStruct = sizeof(ash); //構造体のサイズ ash.pbSrc = bRead; ash.cbSrcLength = cbInBuf; ash.pbDst = (LPBYTE)m_pOutHdr->lpData + (m_cbOutBuf - m_cbOutBufLeft); ash.cbDstLength = m_cbOutBufLeft; if (acmStreamPrepareHeader(m_hAcm, &ash, 0)) break; // 変換 if (acmStreamConvert(m_hAcm, &ash, 0)) break; Preamp((LPBYTE)m_pOutHdr->lpData + (m_cbOutBuf - m_cbOutBufLeft), ash.cbDstLengthUsed); m_cbOutBufLeft -= ash.cbDstLengthUsed; // 後処理 if (acmStreamUnprepareHeader(m_hAcm, &ash, 0)) break; if (m_dwCurrentSize >= m_dwDataSize) return RET_EOF; } return RET_ERROR; } }
static void test_prepareheader(void) { HACMSTREAM has; ADPCMWAVEFORMAT *src; WAVEFORMATEX dst; MMRESULT mr; ACMSTREAMHEADER hdr; BYTE buf[sizeof(WAVEFORMATEX) + 32], pcm[2048], input[512]; ADPCMCOEFSET *coef; src = (ADPCMWAVEFORMAT*)buf; coef = src->aCoef; src->wfx.cbSize = 32; src->wfx.wFormatTag = WAVE_FORMAT_ADPCM; src->wfx.nSamplesPerSec = 22050; src->wfx.wBitsPerSample = 4; src->wfx.nChannels = 1; src->wfx.nBlockAlign = 512; src->wfx.nAvgBytesPerSec = 11025; src->wSamplesPerBlock = 0x3f4; src->wNumCoef = 7; coef[0].iCoef1 = 0x0100; coef[0].iCoef2 = 0x0000; coef[1].iCoef1 = 0x0200; coef[1].iCoef2 = 0xff00; coef[2].iCoef1 = 0x0000; coef[2].iCoef2 = 0x0000; coef[3].iCoef1 = 0x00c0; coef[3].iCoef2 = 0x0040; coef[4].iCoef1 = 0x00f0; coef[4].iCoef2 = 0x0000; coef[5].iCoef1 = 0x01cc; coef[5].iCoef2 = 0xff30; coef[6].iCoef1 = 0x0188; coef[6].iCoef2 = 0xff18; dst.cbSize = 0; dst.wFormatTag = WAVE_FORMAT_PCM; dst.nSamplesPerSec = 22050; dst.wBitsPerSample = 8; dst.nChannels = 1; dst.nBlockAlign = dst.wBitsPerSample * dst.nChannels / 8; dst.nAvgBytesPerSec = dst.nSamplesPerSec * dst.nBlockAlign; mr = acmStreamOpen(&has, NULL, (WAVEFORMATEX*)src, &dst, NULL, 0, 0, 0); ok(mr == MMSYSERR_NOERROR, "open failed: 0x%x\n", mr); memset(input, 0, sizeof(input)); memset(&hdr, 0, sizeof(hdr)); hdr.cbStruct = sizeof(hdr); hdr.pbSrc = input; hdr.cbSrcLength = sizeof(input); hdr.pbDst = pcm; hdr.cbDstLength = sizeof(pcm); mr = acmStreamPrepareHeader(has, &hdr, 0); ok(mr == MMSYSERR_NOERROR, "prepare failed: 0x%x\n", mr); ok(hdr.fdwStatus == ACMSTREAMHEADER_STATUSF_PREPARED, "header wasn't prepared: 0x%x\n", hdr.fdwStatus); mr = acmStreamUnprepareHeader(has, &hdr, 0); ok(mr == MMSYSERR_NOERROR, "unprepare failed: 0x%x\n", mr); ok(hdr.fdwStatus == 0, "header wasn't unprepared: 0x%x\n", hdr.fdwStatus); memset(&hdr, 0, sizeof(hdr)); hdr.cbStruct = sizeof(hdr); hdr.pbSrc = input; hdr.cbSrcLength = sizeof(input); hdr.pbDst = pcm; hdr.cbDstLength = sizeof(pcm); hdr.fdwStatus = ACMSTREAMHEADER_STATUSF_DONE; mr = acmStreamPrepareHeader(has, &hdr, 0); ok(mr == MMSYSERR_NOERROR, "prepare failed: 0x%x\n", mr); ok(hdr.fdwStatus == (ACMSTREAMHEADER_STATUSF_PREPARED | ACMSTREAMHEADER_STATUSF_DONE), "header wasn't prepared: 0x%x\n", hdr.fdwStatus); hdr.cbSrcLengthUsed = 12345; hdr.cbDstLengthUsed = 12345; hdr.fdwStatus &= ~ACMSTREAMHEADER_STATUSF_DONE; mr = acmStreamConvert(has, &hdr, ACM_STREAMCONVERTF_BLOCKALIGN); ok(mr == MMSYSERR_NOERROR, "convert failed: 0x%x\n", mr); ok(hdr.fdwStatus & ACMSTREAMHEADER_STATUSF_DONE, "conversion was not done: 0x%x\n", hdr.fdwStatus); ok(hdr.cbSrcLengthUsed == hdr.cbSrcLength, "expected %d, got %d\n", hdr.cbSrcLength, hdr.cbSrcLengthUsed); todo_wine ok(hdr.cbDstLengthUsed == 1010, "expected 1010, got %d\n", hdr.cbDstLengthUsed); mr = acmStreamUnprepareHeader(has, &hdr, 0); ok(mr == MMSYSERR_NOERROR, "unprepare failed: 0x%x\n", mr); ok(hdr.fdwStatus == ACMSTREAMHEADER_STATUSF_DONE, "header wasn't unprepared: 0x%x\n", hdr.fdwStatus); /* The 2 next tests are related to Lost Horizon (bug 24723) */ memset(&hdr, 0, sizeof(hdr)); hdr.cbStruct = sizeof(hdr); hdr.pbSrc = input; hdr.cbSrcLength = sizeof(input); hdr.pbDst = pcm; hdr.cbDstLength = -4; mr = acmStreamPrepareHeader(has, &hdr, 0); if (sizeof(void *) == 4) /* 64 bit fails on this test */ { ok(mr == MMSYSERR_NOERROR, "prepare failed: 0x%x\n", mr); ok(hdr.fdwStatus == ACMSTREAMHEADER_STATUSF_PREPARED, "header wasn't prepared: 0x%x\n", hdr.fdwStatus); hdr.cbSrcLengthUsed = 12345; hdr.cbDstLengthUsed = 12345; hdr.fdwStatus &= ~ACMSTREAMHEADER_STATUSF_DONE; mr = acmStreamConvert(has, &hdr, ACM_STREAMCONVERTF_BLOCKALIGN); ok(mr == MMSYSERR_NOERROR, "convert failed: 0x%x\n", mr); ok(hdr.fdwStatus & ACMSTREAMHEADER_STATUSF_DONE, "conversion was not done: 0x%x\n", hdr.fdwStatus); ok(hdr.cbSrcLengthUsed == hdr.cbSrcLength, "expected %d, got %d\n", hdr.cbSrcLength, hdr.cbSrcLengthUsed); todo_wine ok(hdr.cbDstLengthUsed == 1010, "expected 1010, got %d\n", hdr.cbDstLengthUsed); mr = acmStreamUnprepareHeader(has, &hdr, 0); ok(mr == MMSYSERR_NOERROR, "unprepare failed: 0x%x\n", mr); ok(hdr.fdwStatus == ACMSTREAMHEADER_STATUSF_DONE, "header wasn't unprepared: 0x%x\n", hdr.fdwStatus); } else ok(mr == MMSYSERR_INVALPARAM, "expected 11, got %d\n", mr); memset(&hdr, 0, sizeof(hdr)); hdr.cbStruct = sizeof(hdr); hdr.pbSrc = input; hdr.cbSrcLength = 24; hdr.pbDst = pcm; hdr.cbDstLength = -4; mr = acmStreamPrepareHeader(has, &hdr, 0); todo_wine { ok(mr == ACMERR_NOTPOSSIBLE, "expected 0x200, got 0x%x\n", mr); ok(hdr.fdwStatus == 0, "expected 0, got 0x%x\n", hdr.fdwStatus); hdr.cbSrcLengthUsed = 12345; hdr.cbDstLengthUsed = 12345; mr = acmStreamConvert(has, &hdr, ACM_STREAMCONVERTF_BLOCKALIGN); ok(mr == ACMERR_UNPREPARED, "expected 0x202, got 0x%x\n", mr); ok(hdr.cbSrcLengthUsed == 12345, "expected 12345, got %d\n", hdr.cbSrcLengthUsed); ok(hdr.cbDstLengthUsed == 12345, "expected 12345, got %d\n", hdr.cbDstLengthUsed); mr = acmStreamUnprepareHeader(has, &hdr, 0); ok(mr == ACMERR_UNPREPARED, "expected 0x202, got 0x%x\n", mr); } /* Less output space than required */ memset(&hdr, 0, sizeof(hdr)); hdr.cbStruct = sizeof(hdr); hdr.pbSrc = input; hdr.cbSrcLength = sizeof(input); hdr.pbDst = pcm; hdr.cbDstLength = 32; mr = acmStreamPrepareHeader(has, &hdr, 0); ok(mr == MMSYSERR_NOERROR, "prepare failed: 0x%x\n", mr); ok(hdr.fdwStatus == ACMSTREAMHEADER_STATUSF_PREPARED, "header wasn't prepared: 0x%x\n", hdr.fdwStatus); hdr.cbSrcLengthUsed = 12345; hdr.cbDstLengthUsed = 12345; hdr.fdwStatus &= ~ACMSTREAMHEADER_STATUSF_DONE; mr = acmStreamConvert(has, &hdr, ACM_STREAMCONVERTF_BLOCKALIGN); ok(mr == MMSYSERR_NOERROR, "convert failed: 0x%x\n", mr); ok(hdr.fdwStatus & ACMSTREAMHEADER_STATUSF_DONE, "conversion was not done: 0x%x\n", hdr.fdwStatus); todo_wine ok(hdr.cbSrcLengthUsed == hdr.cbSrcLength, "expected %d, got %d\n", hdr.cbSrcLength, hdr.cbSrcLengthUsed); todo_wine ok(hdr.cbDstLengthUsed == hdr.cbDstLength, "expected %d, got %d\n", hdr.cbDstLength, hdr.cbDstLengthUsed); mr = acmStreamUnprepareHeader(has, &hdr, 0); ok(mr == MMSYSERR_NOERROR, "unprepare failed: 0x%x\n", mr); ok(hdr.fdwStatus == ACMSTREAMHEADER_STATUSF_DONE, "header wasn't unprepared: 0x%x\n", hdr.fdwStatus); mr = acmStreamClose(has, 0); ok(mr == MMSYSERR_NOERROR, "close failed: 0x%x\n", mr); }
static HRESULT WINAPI ACMStream_fnRead(IAVIStream *iface, LONG start, LONG samples, LPVOID buffer, LONG buffersize, LPLONG bytesread, LPLONG samplesread) { IAVIStreamImpl *This = (IAVIStreamImpl *)iface; HRESULT hr; DWORD size; TRACE("(%p,%d,%d,%p,%d,%p,%p)\n", iface, start, samples, buffer, buffersize, bytesread, samplesread); /* clear return parameters if given */ if (bytesread != NULL) *bytesread = 0; if (samplesread != NULL) *samplesread = 0; /* Do we have our compressor? */ if (This->has == NULL) { hr = AVIFILE_OpenCompressor(This); if (FAILED(hr)) return hr; } /* only need to pass through? */ if (This->cbInFormat == This->cbOutFormat && memcmp(This->lpInFormat, This->lpOutFormat, This->cbInFormat) == 0) { return IAVIStream_Read(This->pStream, start, samples, buffer, buffersize, bytesread, samplesread); } /* read as much as fit? */ if (samples == -1) samples = buffersize / This->lpOutFormat->nBlockAlign; /* limit to buffersize */ if (samples * This->lpOutFormat->nBlockAlign > buffersize) samples = buffersize / This->lpOutFormat->nBlockAlign; /* only return needed size? */ if (buffer == NULL || buffersize <= 0 || samples == 0) { if (bytesread == NULL && samplesread == NULL) return AVIERR_BADPARAM; if (bytesread != NULL) *bytesread = samples * This->lpOutFormat->nBlockAlign; if (samplesread != NULL) *samplesread = samples; return AVIERR_OK; } /* map our positions to pStream positions */ CONVERT_THIS_to_STREAM(&start); /* our needed internal buffersize */ size = samples * This->lpInFormat->nBlockAlign; /* Need to free destination buffer used for writing? */ if (This->acmStreamHdr.pbDst != NULL) { HeapFree(GetProcessHeap(), 0, This->acmStreamHdr.pbDst); This->acmStreamHdr.pbDst = NULL; This->acmStreamHdr.dwDstUser = 0; } /* need bigger source buffer? */ if (This->acmStreamHdr.pbSrc == NULL || This->acmStreamHdr.dwSrcUser < size) { if (This->acmStreamHdr.pbSrc == NULL) This->acmStreamHdr.pbSrc = HeapAlloc(GetProcessHeap(), 0, size); else This->acmStreamHdr.pbSrc = HeapReAlloc(GetProcessHeap(), 0, This->acmStreamHdr.pbSrc, size); if (This->acmStreamHdr.pbSrc == NULL) return AVIERR_MEMORY; This->acmStreamHdr.dwSrcUser = size; } This->acmStreamHdr.cbStruct = sizeof(This->acmStreamHdr); This->acmStreamHdr.cbSrcLengthUsed = 0; This->acmStreamHdr.cbDstLengthUsed = 0; This->acmStreamHdr.cbSrcLength = size; /* read source data */ hr = IAVIStream_Read(This->pStream, start, -1, This->acmStreamHdr.pbSrc, This->acmStreamHdr.cbSrcLength, (LONG *)&This->acmStreamHdr.cbSrcLength, NULL); if (FAILED(hr) || This->acmStreamHdr.cbSrcLength == 0) return hr; /* need to prepare stream? */ This->acmStreamHdr.pbDst = buffer; This->acmStreamHdr.cbDstLength = buffersize; if ((This->acmStreamHdr.fdwStatus & ACMSTREAMHEADER_STATUSF_PREPARED) == 0) { if (acmStreamPrepareHeader(This->has, &This->acmStreamHdr, 0) != S_OK) { This->acmStreamHdr.pbDst = NULL; This->acmStreamHdr.cbDstLength = 0; return AVIERR_COMPRESSOR; } } /* now do the conversion */ /* FIXME: use ACM_CONVERTF_* flags */ if (acmStreamConvert(This->has, &This->acmStreamHdr, 0) != S_OK) hr = AVIERR_COMPRESSOR; This->acmStreamHdr.pbDst = NULL; This->acmStreamHdr.cbDstLength = 0; /* fill out return parameters if given */ if (bytesread != NULL) *bytesread = This->acmStreamHdr.cbDstLengthUsed; if (samplesread != NULL) *samplesread = This->acmStreamHdr.cbDstLengthUsed / This->lpOutFormat->nBlockAlign; return hr; }
BOOL ClSoundDS::AddSoundBuffer(int handle,char *fileBuf,BOOL b3DSound) { HRESULT hr; RiffHead riffHead; FmtHead fmtHead; PCMWAVEFORMAT pcmFormat; LPBYTE lpBlockAdd1, lpBlockAdd2; DWORD blockSize1, blockSize2; CopyMemory(&riffHead,fileBuf,sizeof(RiffHead)); fileBuf += sizeof(RiffHead); if(strncmp(riffHead.riff,"RIFF",4) || strncmp(riffHead.type,"WAVE",4)){ return FALSE; } if(NULL==lpDSBufferTop){ lpDSBufferTop = lpDSBufferTail = new ClSoundBuffer; }else{ lpDSBufferTail->lpNext = new ClSoundBuffer; lpDSBufferTail->lpNext->lpPrev = lpDSBufferTail; lpDSBufferTail = lpDSBufferTail->lpNext; } lpDSBufferTail->bufType = memory_sound; CopyMemory(&fmtHead,fileBuf,sizeof(FmtHead)); fileBuf += sizeof(FmtHead); CopyMemory(&pcmFormat,fileBuf,sizeof(PCMWAVEFORMAT)); if(pcmFormat.wf.wFormatTag == WAVE_FORMAT_PCM){ fileBuf += fmtHead.size; CopyMemory(&fmtHead,fileBuf,sizeof(FmtHead)); fileBuf += sizeof(FmtHead); while(0!=strncmp(fmtHead.fmt,"data",4)){ fileBuf += fmtHead.size; CopyMemory(&fmtHead,fileBuf,sizeof(FmtHead)); fileBuf += sizeof(FmtHead); } DSBUFFERDESC dsbdesc; WAVEFORMATEX audioFmt; ZeroMemory(&dsbdesc, sizeof(DSBUFFERDESC)); dsbdesc.dwSize = sizeof(DSBUFFERDESC); dsbdesc.dwFlags = DSBCAPS_STATIC | DSBCAPS_CTRLVOLUME | DSBCAPS_CTRLPOSITIONNOTIFY | DSBCAPS_LOCSOFTWARE | DSBCAPS_GLOBALFOCUS; if(b3DSound) dsbdesc.dwFlags |= (DSBCAPS_CTRL3D | DSBCAPS_MUTE3DATMAXDISTANCE); dsbdesc.dwBufferBytes = fmtHead.size; CopyMemory(&audioFmt,&pcmFormat,sizeof(WAVEFORMAT)); audioFmt.wBitsPerSample = pcmFormat.wBitsPerSample; audioFmt.cbSize = 0; dsbdesc.lpwfxFormat = &audioFmt; hr = lpDSound->CreateSoundBuffer(&dsbdesc, &lpTmpBuffer, NULL); lpTmpBuffer->QueryInterface(IID_IDirectSoundBuffer8,(LPVOID *)&lpDSBufferTail->lpDSBuffer); RELEASE(lpTmpBuffer); hr = lpDSBufferTail->lpDSBuffer->Lock(0,fmtHead.size, (LPVOID*)&lpBlockAdd1, &blockSize1,(LPVOID*)&lpBlockAdd2, &blockSize2, 0); if(DS_OK==hr){ CopyMemory(lpBlockAdd1,fileBuf,blockSize1); if(fmtHead.size>blockSize1){ CopyMemory(lpBlockAdd2,fileBuf+blockSize1,blockSize2); } lpDSBufferTail->lpDSBuffer->Unlock(lpBlockAdd1, blockSize1, lpBlockAdd2, blockSize2); } lpDSBufferTail->handle = handle; }else{ WAVEFORMATEX dstWF; WAVEFORMATEX *pwfxInfo; WORD cbExtraAlloc; MMRESULT mmResult; fileBuf += sizeof(PCMWAVEFORMAT); CopyMemory(&cbExtraAlloc,fileBuf,sizeof(WORD)); fileBuf += sizeof(WORD); pwfxInfo = (WAVEFORMATEX *)cl_malloc(sizeof(WAVEFORMATEX) +cbExtraAlloc); CopyMemory(pwfxInfo,&pcmFormat,sizeof(PCMWAVEFORMAT)); pwfxInfo->cbSize = cbExtraAlloc; CopyMemory((char *)pwfxInfo +sizeof(WAVEFORMATEX),fileBuf,cbExtraAlloc); fileBuf += (fmtHead.size -sizeof(WAVEFORMATEX)); CopyMemory(&fmtHead,fileBuf,sizeof(FmtHead)); fileBuf += sizeof(FmtHead); while(0!=strncmp(fmtHead.fmt,"data",4)){ fileBuf += fmtHead.size; CopyMemory(&fmtHead,fileBuf,sizeof(FmtHead)); fileBuf += sizeof(FmtHead); } ZeroMemory(&dstWF,sizeof(dstWF)); dstWF.wFormatTag = WAVE_FORMAT_PCM; mmResult = acmFormatSuggest( NULL, pwfxInfo, &dstWF, sizeof(dstWF), ACM_FORMATSUGGESTF_WFORMATTAG); if(mmResult != 0){ myOutputDebugString("オーディオ圧縮が無効です\n"); return FALSE; } HACMSTREAM hAcm = NULL; ACMSTREAMHEADER ash; AcmBuffer acmDst; AcmBuffer acmSrc; acmStreamOpen(&hAcm, NULL,pwfxInfo,&dstWF, NULL, 0L, 0L, ACM_STREAMOPENF_NONREALTIME); acmSrc.dwStreamSize = fmtHead.size; acmStreamSize(hAcm, acmSrc.dwStreamSize, &acmDst.dwStreamSize, ACM_STREAMSIZEF_SOURCE); cl_free(pwfxInfo); acmSrc.lpStream = (LPBYTE)cl_malloc(acmSrc.dwStreamSize); acmDst.lpStream = (LPBYTE)cl_malloc(acmDst.dwStreamSize); ZeroMemory(&ash,sizeof(ash)); ash.cbStruct = sizeof(ash); ash.pbSrc = acmSrc.lpStream; ash.cbSrcLength = acmSrc.dwStreamSize; ash.dwSrcUser = acmSrc.dwStreamSize; ash.pbDst = acmDst.lpStream; ash.cbDstLength = acmDst.dwStreamSize; ash.dwDstUser = acmDst.dwStreamSize; acmStreamPrepareHeader(hAcm,&ash,0); CopyMemory(acmSrc.lpStream,fileBuf,acmSrc.dwStreamSize); MMRESULT nError = acmStreamConvert(hAcm,&ash, ACM_STREAMCONVERTF_BLOCKALIGN); DSBUFFERDESC dsbdesc; ZeroMemory(&dsbdesc, sizeof(DSBUFFERDESC)); dsbdesc.dwSize = sizeof(DSBUFFERDESC); dsbdesc.dwFlags = DSBCAPS_STATIC | DSBCAPS_CTRLVOLUME | DSBCAPS_CTRLPOSITIONNOTIFY | DSBCAPS_LOCSOFTWARE | DSBCAPS_GLOBALFOCUS; if(b3DSound) dsbdesc.dwFlags |= (DSBCAPS_CTRL3D | DSBCAPS_MUTE3DATMAXDISTANCE); dsbdesc.dwBufferBytes = ash.cbDstLengthUsed; dsbdesc.lpwfxFormat = &dstWF; hr = lpDSound->CreateSoundBuffer(&dsbdesc, &lpTmpBuffer, NULL); lpTmpBuffer->QueryInterface(IID_IDirectSoundBuffer8,(LPVOID *)&lpDSBufferTail->lpDSBuffer); RELEASE(lpTmpBuffer); hr = lpDSBufferTail->lpDSBuffer->Lock(0,ash.cbDstLengthUsed, (LPVOID*)&lpBlockAdd1, &blockSize1,(LPVOID*)&lpBlockAdd2, &blockSize2, 0); if(DS_OK==hr){ if(ash.cbDstLengthUsed < blockSize1){ CopyMemory(lpBlockAdd1, acmDst.lpStream, ash.cbDstLengthUsed); ZeroMemory(lpBlockAdd1 +ash.cbDstLengthUsed, blockSize1 -ash.cbDstLengthUsed); if(blockSize2)ZeroMemory(lpBlockAdd2,blockSize2); }else{ CopyMemory(lpBlockAdd1, acmDst.lpStream, blockSize1); if(blockSize2){ CopyMemory(lpBlockAdd2, acmDst.lpStream+blockSize1, ash.cbDstLengthUsed -blockSize1); } } lpDSBufferTail->lpDSBuffer->Unlock(lpBlockAdd1, blockSize1, lpBlockAdd2, blockSize2); } if(hAcm){ acmStreamUnprepareHeader(hAcm,&ash,0); acmStreamClose(hAcm,0); cl_free(acmSrc.lpStream); cl_free(acmDst.lpStream); } lpDSBufferTail->handle = handle; } return TRUE; } // ClSoundDS::AddSoundBuffer
static HRESULT ACMWrapper_ProcessSampleData(TransformFilterImpl* pTransformFilter, LPBYTE data, DWORD size) { ACMWrapperImpl* This = (ACMWrapperImpl*)pTransformFilter; AM_MEDIA_TYPE amt; HRESULT hr; IMediaSample* pSample = NULL; DWORD cbDstStream; LPBYTE pbDstStream; ACMSTREAMHEADER ash; DWORD offset = 0; BOOL stop = FALSE; BOOL unprepare_header = FALSE; MMRESULT res; TRACE("(%p)->(%p,%ld)\n", This, data, size); hr = IPin_ConnectionMediaType(This->tf.ppPins[0], &amt); if (FAILED(hr)) { ERR("Unable to retrieve media type\n"); goto error; } while(!stop) { DWORD rem_buf = This->max_size - This->current_size; DWORD rem_smp = size - offset; DWORD copy_size = min(rem_buf, rem_smp); memcpy(This->buffer + This->current_size, data + offset, copy_size); This->current_size += copy_size; offset += copy_size; if (offset == size) stop = TRUE; if (This->current_size < This->max_size) break; hr = OutputPin_GetDeliveryBuffer((OutputPin*)This->tf.ppPins[1], &pSample, NULL, NULL, 0); if (FAILED(hr)) { ERR("Unable to get delivery buffer (%lx)\n", hr); goto error; } hr = IMediaSample_SetActualDataLength(pSample, 0); assert(hr == S_OK); hr = IMediaSample_GetPointer(pSample, &pbDstStream); if (FAILED(hr)) { ERR("Unable to get pointer to buffer (%lx)\n", hr); goto error; } cbDstStream = IMediaSample_GetSize(pSample); ash.cbStruct = sizeof(ash); ash.fdwStatus = 0; ash.dwUser = 0; ash.pbSrc = This->buffer; ash.cbSrcLength = This->current_size; ash.pbDst = pbDstStream; ash.cbDstLength = cbDstStream; if ((res = acmStreamPrepareHeader(This->has, &ash, 0))) { ERR("Cannot prepare header %d\n", res); goto error; } unprepare_header = TRUE; if ((res = acmStreamConvert(This->has, &ash, This->reinit_codec ? ACM_STREAMCONVERTF_START : 0))) { ERR("Cannot convert data header %d\n", res); goto error; } This->reinit_codec = FALSE; TRACE("used in %lu, used out %lu\n", ash.cbSrcLengthUsed, ash.cbDstLengthUsed); hr = IMediaSample_SetActualDataLength(pSample, ash.cbDstLengthUsed); assert(hr == S_OK); if (ash.cbSrcLengthUsed < ash.cbSrcLength) { This->current_size = ash.cbSrcLength - ash.cbSrcLengthUsed; memmove(This->buffer, This->buffer + ash.cbSrcLengthUsed, This->current_size); } else This->current_size = 0; hr = OutputPin_SendSample((OutputPin*)This->tf.ppPins[1], pSample); if (hr != S_OK && hr != VFW_E_NOT_CONNECTED) { ERR("Error sending sample (%lx)\n", hr); goto error; } error: if (unprepare_header && (res = acmStreamUnprepareHeader(This->has, &ash, 0))) ERR("Cannot unprepare header %d\n", res); if (pSample) IMediaSample_Release(pSample); } return hr; }