Author: carlygale17

Getting to grip with Adobe Media Encoder

Gets videos exporting quickly and to the right quality. This has been a lifesaver saving time and effort getting videos to render properly and easily ready for upload!!

Advertisements

Myspeak.js

meSpeak.loadConfig(“mespeak_config.json”); meSpeak.loadVoice(‘en-us.json’); meSpeak.speak(‘hello world’); meSpeak.speak(‘hello world’, { option1: value1, option2: value2 .. }); meSpeak.speak(‘hello world’, { option1: value1, option2: value2 .. }, myCallback); var id = meSpeak.speak(‘hello world’); meSpeak.stop(id); meSpeak.speak( text [, { option1: value1, option2: value2 .. } [, callback ]] ); text: The string of text to be spoken. The text may contain line-breaks (“\n”) and special characters. Default text-encoding is UTF-8 (see the option “utf16” for other). options (eSpeak command-options): * amplitude: How loud the voice will be (default: 100) * pitch: The voice pitch (default: 50) * speed: The speed at which to talk (words per minute) (default: 175) * voice: Which voice to use (default: last voice loaded or defaultVoice, see below) * wordgap: Additional gap between words in 10 ms units (default: 0) * variant: One of the variants to be found in the eSpeak-directory “~/espeak-data/voices/!v” Variants add some effects to the normally plain voice, e.g. notably a female tone. Valid values are: “f1”, “f2”, “f3”, “f4”, “f5” for female voices “m1”, “m2”, “m3”, “m4”, “m5”, “m6, “m7” for male voices “croak”, “klatt”, “klatt2”, “klatt3”, “whisper”, “whisperf” for other effects. (Using eSpeak, these would be appended to the “-v” option by “+” and the value.) Note: Try “f2” or “f5” for a female voice. * linebreak: (Number) Line-break length, default value: 0. * capitals: (Number) Indicate words which begin with capital letters. 1: Use a click sound to indicate when a word starts with a capital letter, or double click if word is all capitals. 2: Speak the word “capital” before a word which begins with a capital letter. Other values: Increases the pitch for words which begin with a capital letter. The greater the value, the greater the increase in pitch. (eg.: 20) * punct: (Boolean or String) Speaks the names of punctuation characters when they are encountered in the text. If a string of characters is supplied, then only those listed punctuation characters are spoken, eg. { “punct”: “.,;?” }. * nostop: (Boolean) Removes the end-of-sentence pause which normally occurs at the end of the text. * utf16: (Boolean) Indicates that the input is UTF-16, default: UTF-8. * ssml: (Boolean) Indicates that the text contains SSML (Speech Synthesis Markup Language) tags or other XML tags. (A small set of HTML is supported too.) further options (meSpeak.js specific): * volume: Volume relative to the global volume (number, 0..1, default: 1) Note: the relative volume has no effect on the export using option ‘rawdata’. * rawdata: Do not play, return data only. The type of the returned data is derived from the value (case-insensitive) of ‘rawdata’: – ‘base64‘: returns a base64-encoded string. – ‘mime‘: returns a base64-encoded data-url (including the MIME-header). (synonyms: ‘data-url’, ‘data-uri’, ‘dataurl’, ‘datauri’) – ‘array‘: returns a plain Array object with uint 8 bit data. – default (any other value): returns the generated wav-file as an ArrayBuffer (8-bit unsigned). Note: The value of ‘rawdata’ must evaluate to boolean ‘true’ in order to be recognized. * log: (Boolean) Logs the compiled eSpeak-command to the JS-console. callback: An optional callback function to be called after the sound output ended. The callback will be called with a single boolean argument indicating success. If the resulting sound is stopped by meSpeak.stop(), the success-flag will be set to false. Returns: * if called with option rawdata: a stream in the requested format (or null, if the required resources have not loaded yet). * default: a 32bit integer ID greater than 0 (or 0 on failure). The ID may be used to stop this sound by calling meSpeak.stop(<id>). if (meSpeak.isVoiceLoaded(‘de’)) meSpeak.setDefaultVoice(‘de’); // note: the default voice is always the the last voice loaded meSpeak.loadVoice(‘fr.json’, userCallback); // userCallback is an optional callback-handler. The callback will receive two arguments: // * a boolean flag for success // * either the id of the voice, or a reason for errors (‘network error’, ‘data error’, ‘file error’) alert(meSpeak.getDefaultVoice()); // ‘fr’ if (meSpeak.isConfigLoaded()) meSpeak.speak(‘Configuration data has been loaded.’); // note: any calls to speak() will be deferred, if no valid config-data has been loaded yet. meSpeak.setVolume(0.5); meSpeak.setVolume( volume [, id-list] ); Sets a volume level (0 <= v <= 1) * if called with a single argument, the method sets the global playback-volume, any sounds currently playing will be updated immediately with respect to their relative volume (if specified). * if called with more than a single argument, the method will set and adjust the relative volume of the sound(s) with corresponding ID(s). Returns: the volume provided. alert(meSpeak.getVolume()); // 0.5 meSpeak.getVolume( [id] ); Returns a volume level (0 <= v <= 1) * if called without an argument, the method returns the global playback-volume. * if called with an argument, the method will return the relative volume of the sound with the ID corresponding to the first argument. if no sound with a corresponding ID is found, the method will return ‘undefined’. var browserCanPlayWavFiles = meSpeak.canPlay(); // test for compatibility // export speech-data as a stream (no playback): var myUint8Array = meSpeak.speak(‘hello world’, { ‘rawdata’: true }); // typed array var base64String = meSpeak.speak(‘hello world’, { ‘rawdata’: ‘base64’ }); var myDataUrl = meSpeak.speak(‘hello world’, { ‘rawdata’: ‘data-url’ }); var myArray = meSpeak.speak(‘hello world’, { ‘rawdata’: ‘array’ }); // simple array // playing cached streams (any of the export formats): meSpeak.play( stream [, relativeVolume [, callback]] ); var stream1 = meSpeak.speak(‘hello world’, { ‘rawdata’: true }); var stream2 = meSpeak.speak(‘hello again’, { ‘rawdata’: true }); var stream3 = meSpeak.speak(‘hello yet again’, { ‘rawdata’: ‘data-url’ }); meSpeak.play(stream1); // using global volume meSpeak.play(stream2, 0.75); // 75% of global volume meSpeak.play(stream3); // v.1.4.2: play data-urls or base64-encoded var id = meSpeak.play(stream1); meSpeak.stop(id); Arguments: stream: A stream in any of the formats returned by meSpeak.play() with the “rawdata”-option. volume: (optional) Volume relative to the global volume (number, 0..1, default: 1) callback: (optional) A callback function to be called after the sound output ended. The callback will be called with a single boolean argument indicating success. If the sound is stopped by meSpeak.stop(), the success-flag will be set to false. (See also: meSpeak.speak().) Returns: A 32bit integer ID greater than 0 (or 0 on failure). The ID may be used to stop this sound by calling meSpeak.stop(<id>). meSpeak.stop( [<id-list>] ); Stops the sound(s) specified by the id-list. If called without an argument, all sounds currently playing, processed, or queued are stopped. Any callback(s) associated to the sound(s) will return false as the success-flag. Arguments: id-list: Any number of IDs returned by a call to meSpeak.speak() or meSpeak.play(). Returns: The number (integer) of sounds actually stopped.

Note on export formats, ArrayBuffer (typed array, defaul) vs. simple array:
The ArrayBuffer (8-bit unsigned) provides a stream ready to be played by the Web Audio API (as a value for a BufferSourceNode), while the plain array (JavaScript Array object) may be best for export (e.g. sending the data to Flash via Falsh’s ExternalInterface). The default raw format (ArrayBuffer) is the preferred format for caching streams to be played later by meSpeak by calling meSpeak.play(), since it provides the least overhead in processing.

meSpeak.speakMultipart() — concatenating multiple voices

Using meSpeak.speakMultipart() you may mix multiple parts into a single utterance.

See the Multipart-Example for a demo.

The general form of meSpeak.speakMultipart() is analogous to meSpeak.speak(), but with an array of objects (the parts to be spoken) as the first argument (rather than a single text):

meSpeak.speakMultipart( <parts-array> [, <options-object> [, <callback-function> ]] ); meSpeak.speakMultipart( [ { text: “text-1”, <other options> ] }, { text: “text-2”, <other options> ] }, … { text: “text-n”, <other options> ] }, ], { option1: value1, option2: value2 .. }, callback );

Only the the first argument is mandatory, any further arguments are optional.
The parts-array must contain a single element (of type object) at least.
For any other options refer to meSpeak.speak(). Any options supplied as the second argument will be used as defaults for the individual parts. (Same options provided with the individual parts will override these defaults.)
The method returns — like meSpeak.speak() — either an ID, or, if called with the "rawdata" option (in the general options / second argument), a stream-buffer representing the generated wav-file.

Note on iOS Limitations

iOS (currently supported only using Safari) provides a single audio-slot, playing only one sound at a time.
Thus, any concurrent calls to meSpeak.speak() or meSpeak.play() will stop any other sound playing.
Further, iOS reserves volume control to the user exclusively. Any attempt to change the volume by a script will remain without effect.
Please note that you still need a user-interaction at the very beginning of the chain of events in order to have a sound played by iOS.

Note on Options

The first set of options listed above corresponds directly to options of the espeak command. For details see the eSpeak command documentation.
The meSpeak.js-options and their espeak-counterparts are (mespeak.speak() accepts both sets, but prefers the long form):

meSpeak.js eSpeak
amplitude -a
wordgap -g
pitch -p
speed -s
voice -v
variant -v<voice>+<variant>
utf16 -b 4 (default: -b 1)
linebreak -l
capitals -k
nostop -z
ssml -m
punct –punct[=”<characters>”]

Voices Currently Available

  • ca (Catalan)
  • cs (Czech)
  • de (German)
  • el (Greek)
  • en/en (English)
  • en/en-n (English, regional)
  • en/en-rp (English, regional)
  • en/en-sc (English, Scottish)
  • en/en-us (English, US)
  • en/en-wm (English, regional)
  • eo (Esperanto)
  • es (Spanish)
  • es-la (Spanish, Latin America)
  • fi (Finnish)
  • fr (French)
  • hu (Hungarian)
  • it (Italian)
  • kn (Kannada)
  • la (Latin)
  • lv (Latvian)
  • nl (Dutch)
  • pl (Polish)
  • pt (Portuguese, Brazil)
  • pt-pt (Portuguese, European)
  • ro (Romanian)
  • sk (Slovak)
  • sv (Swedish)
  • tr (Turkish)
  • zh (Mandarin Chinese, Pinyin)*
  • zh-yue (Cantonese Chinese, Provisional)**

References

O’Keefe, G., Clarke-Pearson, k., 2011. The Impact of Social Media on Children, Adolescents and Familes. Available from http://pediatrics.aappublications.org/content/127/4/800.full.pdf+html [Accessed 7May 2014]

Dial, M., 2013. Who to follow on Twitter? Types of Twitter Profiles and Users. Social Media Today. 1 October 2013. Available from http://socialmediatoday.com/minterdial/1781106/who-follow-twitter-cartography-types-twitter-profiles-and-users  [Accessed 7 May 2014]

Elliot, A., 2013. The 9 Types of Facebook User. Social Media Today. Available from http://socialmediatoday.com/socialbarrel/1646241/9-types-facebook-users-infographic. [Accessed on 12 May 2014]

Griggs.B., 2009. The 12 most annoying types of Facebookers. CNN.com/technology. Available from http://edition.cnn.com/2009/TECH/08/20/annoying.facebook.updaters/ [Accessed on 2May 2014]

Bradford School of Management. 2014. Introduction to research and research methods. Available from http://www.brad.ac.uk/management/media/management/els/Introduction-to-Research-and-Research-Methods.pdf [Accessed 4 Feb 2014]

Crawford 2013. http://ptgmedia.pearsoncmg.com/images/9780321864970/samplepages/0321864972.pdf

Eisele, R., 2008-2014. JQuery Webecam Plugin

http://www.xarg.org/project/jquery-webcam-plugin/

MeSpeak Java Plugin. 2014. http://www.masswerk.at/mespeak/

Kelley, L., Ugenheimer, D., 2008. Advertising Media Planning. Ney York. ME Sharpe Inc

Allen. 2011. Artist Management for the Music Business 2e

Emerson, S., Wortham E., 2002. Narratives in Action: A Strategy for Research and Analysis

Wilson 2002. Information Science and Research Methods. Available from http://www.informationr.net/tdw/publ/papers/slovak02.html. [Accessed 12 May 2014]

Wallop, H. 2013. 12 Signs you are already addicted. The telegraph online. 12 Sept 2013. Available from http://www.amplifon.co.uk/resources/impact-of-sound-on-the-brain/ [Accessed 9 May 2013]

Stopa, L., Clark, D., 2003. Behaviour and Research Therapy. http://www.sciencedirect.com/science/article/pii/000579679390024O

Roland Barthes 1968

Sausaurre 1974

Sosnoski 2010

Gerbner 1960

BBC Academy Transmedia Story Telling 101

Cosmo 2014

To Conclude…

The graduate project has been a very insightful learning experience. I have testing my theoretical mind, my technological skill, my creative flair, my patience and my confidence. The skill set I am most proud of developing is my knowledge of new softwares, After Effects and Premiere Pro. Having self taught myself from start to finish I feel I have come a very long way very quickly. I thoroughly enjoyed filming and editing and this has even opened a few doors in terms of career path ideas. I have extended my knowledge of Photoshop and Illustrator, which will also help with career paths. Having struggled with code I am proud of what I have achieved even though the project doesn’t look quite like how I wanted it to look I have learned so much and have a new understand of Javascript, JQuery, HTML and CSS. I think what this project lack in visible successes it makes up for in long term personal and professional development.

Although I am relatively happy with the the final outcome of the project I feel there is room for massive improvements to be made. Like I have iterated throughout, I think more content is necessary and the webpage needs to be more interactive, with everything playing at one or not at all or playing here and moving there. It needs to push the ideas further, to stretch this online meets offline conceptualisation. I think the 3D gallery scroller is almost like a demo of the real thing and looks a bit unfinished. This has come from making a change last minute – so this is a very very important lesson I have learned. To either make a decision and stick to it and put it into place immediately instead of hoping everything will be okay. In my eyes my time management was good, i.e. I spent a lot of time on the project (all day everyday in the library) but because I did not set weekly targets and mini deadlines for myself I got a bit lost in the workload. I let things that took a long time take even longer because I was a little too casual about it, so this will be a lesson for the future.

In terms of my audience research process I do not think I could have been more thorough, with a minimum of 4 different methods I think I have researched it from all angles. This was the part I have actually enjoyed the most. In terms of the design side, again I should have set personal mini deadlines. By creating the content first and leaving coding the website until quite late I think it looks rushed and amateur. My overall design process of my video and clips I think has been strong. By critiquing them with other people and getting them to be honest I think these came out the most successful. As I have said before, I wish now we were starting the project because I could utilise my skills learned, whereas before it seemed I was wasting time learning them although I did need to start someone. So, although the overall look and feel of the site is not what I wanted and there are many shortfalls, I think that I have shown the strength and dedication to continue on and rise above them to complete a project of substantial scope and depth.

Part of me wishes that I had continued with the second idea of a very graphicsy type interactive joiner image because I think a joiner image is so powerful it just stands alone and proud in its own right. My concerns with this was using people images off the internet, for example although the images I have used from Instagram have all been for public viewing therefore their should be no real issue. Also due to the nature of the video it moves too quick for the photographers identity to be reviewed anyways. This is the idea I would pursue in future whereby the images would be pulled live and joiner together more systematically and incorporated more code so not only being a narrative but more interactive as well whilst letting the data speak for itself.

I feel I have learned a lot about the ideas development process and proved that I carry out a product form start to finish no matter what it is. I think I have showed willingness to be innovative and unique by choosing a subject and treatment that is a bit unusual and not got a set structure you can follow. This shows determination and commitment to the task in hand. I feel I have learned so much in terms of organisation, planning, graphics, video, video editing, sound editing, storyboarding, working with others, coding and designing. So although I am not necessarily happy with the final product, I am happy with what it has taught me for the future.

Digital Story Telling Example

bear 76

Bear 7

A very strong example of digital narrative developed from a tiny seed of thought. I like the voice over element and the continuous movement on the page. This is not just telling the story of one bear, it is telling the conceptual story of what the bear stands for and everything else surround human society and how we live. I like the graphic stats meets real life filming as both offer strong visual contrast from online and offline. The negatives with this piece is that there is a feeling of anti climax, like you are waiting for something major to happen and for all to be revealed but instead it just plays on and on and you never quite get to the end of it. This would put some viewers off and make them switch over straight away which I want to avoid.

Linear to Multi Narrative Design

Linear Narrative

 

 

Multiple Overlapping narrative Design.

 

SITE5

 

So imagine the user is greeted by a strange selection of images. They begin clicking on them and pop ups start happing. People don’t like pop ups… they make them feel uneasy and annoyed. This is how an emotive response will be gained.

SITE3

 

 

So as the view beginning to click and interact, opening and closing the pop ups, they begin to recognise things a little more and begin to make the tie between everything and find the connecting similarity.

SITE2

 

They may decide to close and open them one by one, or open them all at the same time, or leave some playing. The build up of noise and page traffic and general mash up of images that all represent something unique to do with social media not only tell a story in their own right but tell the story as a whole with the bizarre and unnatural layout on the page.

 

 

 

SITE1

Initial 3D layering narrative design ideas. The element of confusion is a good thing – its representative of the complicated lives led online and off. The noise and sounds, and recognisable things will make the user either want to interact to make it stop or interact to make it carry on and to see more. By using the interactive function as well, with the webcam plugin and the speech plugin I think the overall concept will get across and will “be the story”.