04/29/20 – Fanglan Chen – Accelerating Innovation Through Analogy Mining

Summary

Hope’s paper “Accelerating Innovation Through Analogy Mining” studies how to boost knowledge discovery through searching for analogies in massive and unstructured real-world datasets. This research is motivated by the availability of large idea repositories which can be used as databases to search for analogous problems. However, it is very challenging to find useful analogies among the massive and noisy real-world repositories. Manual and automated methods have their own advantages and disadvantages: hand-created databases have a high relational structure which is central to analogy search but expensive to obtain; naive machine learning or information retrieval can be easily scaled to large datasets with similarity metrics but fail to incorporate structural similarity. To address the challenges, the researchers explore the potential of learning “problem schemas,” which are simpler structural representations that specify the purpose and mechanism of a product. Their proposed approach leverages the power of crowdsourcing and recurrent neural networks to extract purpose and mechanism vector representations from product descriptions. The experimental results indicate that the learned vectors can facilitate the search of analogies with higher precision and recall than traditional information retrieval methods. 

Reflection

This paper introduces an innovative approach for analogy search, the task of which is very similar to the “SOLVENT” paper we discussed last week. The “weaker structural representations” idea is very interesting and it allows more flexibility in automatic algorithm design compared with relying on fully structured analogical reasoning. My impression is that the human component in the model design is comparatively weaker than that of the approaches presented in other readings we discussed before. Crowd work in this paper is leveraged as an approach to generate training data and evaluate the experimental results. I have been thinking about if there is any place that has the potential to incorporate human interaction in the model design. As we know, recurrent neural networks have certain limitations. The first weakness is variable length, which means the RNN models cannot handle long sequence data, and this largely constrains the usage scenarios. The second weakness is the sliding window, which ignores the continuity between the sliced subsequences, which is more close to the surface but not as deep as the paper claims. I am wondering if there is a possibility to leverage human interaction to overcome the shortcoming of the model itself.

The performance of machine learning models is highly driven by the quality of the training data. With the crowdsourced product purpose and mechanism annotations, I feel there is a need to incorporate some quality control components in the proposed framework. A few papers we discussed before touching upon this point. Also, though very powerful in numerous complex tasks, deep learning models are usually criticized due to its lack of interpretability. Although the RNN performance reported in the paper in regards to recall and precision is better than that of traditional information retrieval methods. However, those similarity-based methods have their own merits as their mechanism and decision boundaries are more transparent so it would be possible to detect where the problem is and why the results are not desirable. In this case, there is a trade-off between the model performance (accuracy, precision, and recall) and interpretability, it is worthy of thinking about which one to choose over the other.

Discussion

I think the following questions are worthy of further discussion.

  • What do you think can incorporate more human components in the model design?
  • Compared with last week’s approach SOLVENT, which one you think works better? Why?
  • What are some other potential applications for this system outside of knowledge discovery?
  • Do you think the recurrent neural network method is better than traditional similarity-based methods such as TF-IDF in the analogy search task and other NLP tasks? Why or why not?

Read More

04/29/2020 – Ziyao Wang – IdeaHound: Improving large-scale collaborative ideation with crowd-powered real-time semantic modeling

Innovation without guidance is a hard task. To help the users, the authors proposed that a map of the solution space that is being explored can inspire and direct exploration. With research on current automated approaches, they found that all the approaches cannot build adequate maps. In addition to this, all the current deployed crowdsourcing approaches require external workers to do tedious semantic judgment tasks. To resolve this problem, they presented IdeaHound. The system can seamlessly integrate semantic tasks of organizing ideas into users’ idea generation activities. They used several case studies to prove that the system can yield high-quality maps without detracting from idea generation. And the users are found willing to use the system to simultaneously generate and organize ideas, and the system can obtain more accurate models than existing approaches.

Reflections:

The idea of the paper is really interesting. Instead of hiring crowd workers to do the idea clustering, they designed an interface that can let the users write their ideas and make the users able to group the ideas by themselves. Instead of letting a group of people write the ideas and hiring another group of people to cluster the ideas, making the first group of people able to do both idea generation and idea clustering is more efficient and more accurate. According to the idea groups, the system can generate a map with all ideas locates on it. If two groups of ideas have similarities, they will position nearby. With this map, the users can be inspired and provide more ideas. Additionally, if the users write some ideas, the system will automatically recommend some ideas which have similarity with them. This can also inspire users to come up with more novel ideas. I really like the example in the authors’ presentation. When not given an inspiring map or generate ideas automatically, we may get outputs like pizzas with top of broccoli, which is hardly accepted by most of the people. This makes me aware of the importance of the system.

I think this kind of interaction should be learned by all other current applications which deployed crowdsourcing. The users of the systems are willing to do some more tasks than required if they are interesting or they think the tasks are meaningful. Humans are not hired by AIs. Instead, AIs should be the helper of humans when humans are doing some tasks. As is said, we should leverage both human strength and machine strength. In this kind of idea generation task, it is more efficient to let machines support humans to complete the tasks.

Questions:

Will you do the ideas clustering automatically when you are only asked to provide ideas while you are able to do the clustering using the system?

Can this kind of idea, letting users able to do something instead of requiring them to do something, being applied to other applications?

Do you like pizzas with top of broccoli? What is your opinion on these ideas generated by systems? What about using a system to generate ideas and let human workers select the useful ones from all the ideas?

Read More

04/29/2020 – Palakh Mignonne Jude – VisiBlends: A Flexible Workflow for Visual Blends

SUMMARY

In this paper, the authors propose a flexible workflow to enable the creation of visual blends. The authors specifically focus on the ‘Hybrid’ visual metaphor wherein objects are ‘fused together’. The authors propose a blending design pattern ‘Single Shape Mapping’ which can be used to blend two objects with similar shapes and blends all of one object into a part of the other one. The workflow consists of different microtasks including brainstorming, annotation, and evaluation. The entire workflow consists of 6 steps – brainstorming, finding images, annotating images, detecting images that blend together (by the system), automatic synthesizing of the blends, evaluating the blends (by the user) – and the users were made to watch a 15-minute training session before they started with the task. The authors evaluated the task decomposition with three studies namely decentralized collaboration,  group collaboration on blends for messages, and novice users with and without VisiBlends. The authors found that VisiBlends was helpful as it enabled people to meet all of the constraints associated with visual blending.

REFLECTION

I enjoyed reading this paper and found the motivation for this study to be very nice. It was interesting to see how creativity, which is predominantly a human affordance, was being presented in a mixed-initiative setting.  I liked the examples that were chosen throughout the paper and found that they helped me to better understand the challenges associated with blending images (orange + healthy, with apple as the health symbol) as well as to appreciate the images that were generated well (orange + healthy, with a health symbol that was not food).

The study on ‘decentralized collaboration’ reminded me of the paper on ‘The Knowledge Accelerator: Big Picture Thinking in Small Pieces’ which was discussed last week. I liked the study on the ability of novice users to create these visual blends with and without VisiBlends. I also agree that having a flexible iterative workflow similar to the one used in the paper is very useful as it aids the users to identify issues with the original results and then improve upon the same.

I liked that the authors discuss how creative design problems also have patterns in the ‘Discussion’ section. I think it would be very interesting to see a similar study be conducted in the domain of story writing. Text data poses a multitude of challenges and having a decomposed workflow such as this proposed for writing would be very interesting – especially given that authors may have very varied writing styles.

(Interestingly, “Wash your hands. It’s the smart move” was one of the messages used as part of this study.)

QUESTIONS

  1. What modifications can be made to the interface design of the VisiBlends website to better aid users in creating these visual blends? What are the main drawbacks of the existing interface?
  2. The authors propose the decomposed workflow for ‘Hybrid’ visual metaphors. Is it possible to create such a workflow for the ‘Simile’ or ‘Contextual Metaphor’ visual blends? What kind of changes would be required to be made to enable this?
  3. The authors conduct a study to evaluate the usage of this system by novices. What results would have been yielded if the users of the system were people with a background In marketing, but who were not graphic designers? Would their approach to the problem have been different?

Read More

04/29/2020 – Palakh Mignonne Jude – Accelerating Innovation Through Analogy Mining

SUMMARY

In this paper, the authors attempt to facilitate the process of finding analogies with a view to boost creative innovations by exploring the value that can be added by incorporating weak structural representations. They leverage the vast body of online information available (for the purpose of this study, product descriptions from Quirky.com). They generate microtasks for crowdworkers to perform that were designed to label the ‘purpose’ and ‘mechanism’ parts of a product description. The authors use GloVe word vectors to represent their purpose and mechanism words and use a BiRNN to learn the purpose and mechanism. In order to collect analogies, they use AMT crowd workers to find analogies for 8000 product descriptions. In the evaluation stage, the authors attempt to weigh the usefulness of their algorithm by having participants redesign a product. 38 AMT workers were recruited for the same and the task was to design a cell phone charger case. 5 graduate students were recruited to evaluate the ideas generated by the workers. Based on a predefined criterion of ‘good’ ideas, 208 were produced out of 749 total ideas (with 2 judges rating it as good) and 154 were produced out of 749 total ideas (with 3 judges rating it as good). In both cases, the analogy approach proposed by the authors out-performed the TF-IDF baseline model and random model.

REFLECTION

I found the motivation of this study to be very good – especially based on ‘bacteria-slot machine’ analogy example highlighted in the introduction of the paper. I agree that given the vast amount of data available, having such a system that would accelerate the process of finding analogies could very well aid in quicker innovation and discovery.

I like that the authors chose to present their approach by using product descriptions. I also like the use of ‘purpose’ and ‘mechanism’ annotations and feel that given the more general domain of this study, the quality of annotations by the crowdworkers would be better than in the case of the paper on ‘SOLVENT: A Mixed Initiative System for Finding Analogies between Research Papers’.

I also liked that the authors presented the results given by using the TF-IDF baseline as it indicated the findings that would have been generated by near-domain results. I felt that it was good that the authors added a criterion to judge the feasibility of an idea, one that could be implemented using existing technologies.

Additionally, while I found that the study and methods proposed by this paper were good, I did not like the organization of the paper.

QUESTIONS

  1. How would you rate the design of the interface used to collect ‘purpose’ and ‘mechanism’ annotations? What changes might you propose to make this better?
  2. The authors do not mention the details or experiences of the AMT workers. How much would workers prior experience influence their ability to find ideas using this approach? Can this approach aid more experienced people working with product innovations?
  3. The authors of SOLVENT leverage the mixed-initiative system proposed by this paper to find analogies between research papers. Which are the domains where this approach would fail to a great extent (even if modifications were to be made)?

Read More

04/29/20 – Fanglan Chen – DiscoverySpace: Suggesting Actions in Complex Software

Summary

Fraser et al.’s paper “DiscoverySpace: Suggesting Actions in Complex Software” explores how action suggestions can be incorporated into a complex system that can help beginners and inexperienced users navigate and gain confidence while interacting with the system. This research is motivated by the observation that there are several problems that novice users might face several problems when trying to use complex software: (1) new users may be unfamiliar with the vocabulary used in the application, making it difficult to explore the desired features; (2) online tutorials might be difficult to suit their situation and goals; (3) several shortcuts in the software to accomplish the same task are available but inexperienced users might get overwhelmed by the multiple approaches and have no idea how to locate the most efficient ones; (4) users only get exposure to a small number of software features which may limit their cognition of its potential power. To address the above issues, the researchers develop DiscoverySpace which is a prototype action suggestion software extension for Adobe Photoshop to help beginners get started without feeling overwhelmed.

Reflection

I think this paper conducted a very interesting study on how the proposed action recommendations can help novice users build confidence, accomplish tasks, and discover features. Probably many of us have similar experiences that when we are trying to get started with a complex software used by professionals in different domains, the learning curve is so deep that we feel frustrated and discouraged at the beginning. If there is no necessity to use the software or light-weighted alternatives available, we may give up using it. The proposed extension does a good job presenting some basic features of the complex software and develops “one-click” operations for easy tasks, which can make new users gain some sense of achievement while interacting with the system and may want to continue exploring other features.

As presented in the analysis and experimental results, the largest improvement of the proposed Adobe Photoshop extension is among the users who just get started using the software. The evaluation is mostly based on their self-reported confidence. However, another important aspect is ignored in the process – how much did the users learn in the process. That relates to what goal we want to achieve by using professional software. If the users just expect to leverage the power of Photoshop to do really simple tasks, there are a bunch of mobile applications available with only a few easy clicks. If the objective is just to beautify a selfie, the Instagram application has built-in operations and is very easy to use. As we know, Photoshop is used by professionals for image editing. If the users would like to learn how to use the software and build up their skills over time, the current version of the proposed extension does not seem to be helpful. The proposed approach is to encapsulate a sequence of operations into a single click. There is no denying the fact that it is very easy to use, but the light-weighted operations may not contribute to long-term learning. I am a Photoshop user and use it frequently as needed. The current version of the proposed extension may not be very useful to me, but I feel there is a lot of potentials to improve to make it more powerful. Firstly, it would be very useful to have a dialogue box to present the back-end steps conducted within a one-click function. Knowing the basic sequence to achieve a simple task can help the users build their knowledge and know what to do or at least what to search when they need to achieve a similar task. Secondly, it would be helpful to have some interactive modules that enable users to adjust a number of parameters, such as brightness, contrast, and so forth. These are fundamentals for users who want to enhance their skill levels and get experienced with Photoshop.

Discussion

I think the following questions are worthy of further discussion.

  • In what user scenarios you think the proposed software extension would be most useful?
  • Do you think it is helpful to incorporate a sequence of operations in one click or there is a need to present the operations step by step to users?
  • Do you think that this approach with newly released extensions can assist experts in complex professional software?
  • Can you think about some examples of other applications in the same or different domains that could benefit from the proposed approach?

Read More

4/29/20 – Lee Lisle – DiscoverySpace: Suggesting Actions in Complex Software

Summary

               Fraser et al.’s paper discusses how difficult some software programs are, especially for novices, and a possible solution to help ameliorate the steep learning curves they often possess. Their solution is the DiscoverySpace interface – a tool for the creation of macros to help novice users learn new capabilities of the software they want to use. They further test this with a prototype that is used to help participants learn how to use the popular image editing program Photoshop. After finding 115 different macros from various sources, they designed their workflow and interface and ran a study to evaluate it. They found, in a study with 28 participants, that found that participants were significantly less likely to be discouraged or say they couldn’t figure something out with the DiscoverySpace tool installed with Photoshop.

Personal Reflection

               This paper provides a fantastic workflow for easing novice users into using new and difficult programs. I liked it because it provides a slightly more customized experience than the youtube video walkthroughs and online tutorials that I’m accustomed to using. I would even like to actually use this interface for Photoshop, as its one program I attempted to break into a few times early on in my college career, always failing as there were too many features too obliquely described.

               I was surprised that the authors removed the suggestions that contained pauses and dialogue. I would have expected those situations to be better able to present the user with the appropriate background for the effects they wanted to do. However, when they explained their reasoning – that the explanations often were not enough and confused the users – it made a lot more sense to remove those altogether.

               I’m not sure how I feel about their comment that later on they are supporting “paid” actions, where the macros do something that can be considered of a higher quality and thereby require some form of compensation for the macro creator. I don’t think an academic paper is the place for that sort of suggestion, as it doesn’t really add to the software or approach the paper presents. All tools that academic papers present could be used in commercial software, so why would that be of particular note in this paper?

               Lastly, and this is more of a quibble, I was more off put than I thought I would be by the text-aligned images as seen in figures 3 and 4. The alignment is more difficult to read than it would be in a casual magazine-type environment, and should be reserved for that sort of publication.

Questions

  1. Do you think the 115 presented actions are enough of a testbed for the prototype tool? I.E., should they have more to present a better amalgamation of possible uses? How would they generate more?
  2. Beyond using image analysis to present some initial ideas to the users, what other ways might you improve their approach to make it more automated, or do you think there’s enough or too much automation already?
  3. What other programs could use this approach, and how might they integrate it into their platforms?

Read More

04/29/20 – Sukrit Venkatagiri – Accelerating Innovation Through Analogy Mining

Paper: Tom Hope, Joel Chan, Aniket Kittur, and Dafna Shahaf. 2017. Accelerating Innovation Through Analogy Mining. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD ’17), 235–243. https://doi.org/10.1145/3097983.3098038

Summary: This paper talks about the challenge of mining analogies from large, real-world repositories, such as patent databases. Such databases pose challenges because they are highly relational but sparse in nature. This s a reason why machine learning approaches do not fare well when applied to these types of databases, especially since they cannot formulate a patter of the underlying structure, which is important for analogy mining. The corpora are also expensive to build, store, and update, while automation cannot be easily applied. The authors overcome these limitations by leveraging the creativity of the crowd and affordable computational capabilities of RNNs. The approach is a structured purpose-mechanism schema for identifying analogies between two research papers. Finally, the authors evaluate crowd worker performance by asking graduate students to annotate the ideas generated around three main ideas: quality, novelty, and feasibility. They find that their approach increased feasibility among the participants in the study.

Reflection:
Overall, I really liked the paper in how it attempts to solve a hard problem by using a scalable approach: crowds and RNNs, and tests it on a real-world dataset. I also liked how the paper defines similarity between different ideas (i.e. analogies) based on purpose and the mechanisms through which products work. Further, the paper suggests more complex metrics for research papers. This raises the question: how much more difficult is it to mine analogies for complex/more abstract ideas, compared to simple ideas? Perhaps structured labels could help in that regards.

The approach itself is commendable since it is a great example of a mixed-initiative user interface that combines the creativity of the crowd and the affordable computation of RNNs. Further, this approach does not needlessly waste human computation. The authors also completed a thorough evaluation of the machine intelligence portion.

Second, I appreciate the approach taken towards making something subjective—in this case, creativity—into something more objective, by breaking it down into different rate-able metrics.

Finally, the idea of using randomly generate analogies to “spark creativity” and the results of that show that creativity really does need diverse ideas. I wonder why this may be, and how to introduce such randomness into real-world work practice.

Questions:
1. How scalable do you think the system is? What other limitations does it have?
2. Can this approach be used to generate analogies in other fields? What would be different?
3. Do you think creativity is subject? Can it be made into something objective?

                                                                          

Read More

04/29/2020 – Bipasha Banerjee – VisiBlends: A Flexible Workflow for Visual Blends

Summary

The authors define a method to include graphics to draw attention to messages, specifically the blending of two images from two different concepts. They started the paper with the beautiful example of Starbucks and summer, and how the two ideas generated the resultant image. The paper mentions that the process they used was the hybrid blend of two images into one object, which represents both the concepts effectively. The process of visual blending is defined here to be two concepts resulting in two objects and then the integration of both the objects into one, yet, making sure that both the original objects are visible. The VisiBlend system involves various steps in the entire workflow. Firstly, the users need to brainstorm associated concepts, then find images from both the concepts, followed by annotating them correctly. Finally, the algorithm chooses the images to blend and generates final images for the mix, which is evaluated by the user. It was found that the tool helped in the improvement of the resultant image. 

Reflection

This is an interesting study that can be used in the field of advertisement and marketing, among others. The main attraction of the paper is that the system helps people who are novices in graphic designing. The steps are straightforward to follow, and the results showed that using the VisiBlend system proved to be more helpful. This paper essentially tries to find the similarity between images as opposed to texts. 

One of the drawbacks that the users pointed out was the time taken to find relevant images to a concept. I think this problem could be mitigated if we were to even consider scenes as opposed to a single image. Various computer vision algorithms provide tools to get bounding boxes and improve image quality. This would help users save time and get components from larger images too.

I also kept on thinking about how we can extend this idea to other possible areas. One thing that comes to my mind is if we could combine images and text to find analogies between them. By combining them, I do not mean just providing a keyword and getting an image. That is what Google or any search algorithm does. However, I mean, if a user searches for a word, then gets an image they like, could we link articles to that? We could also try and find text related to the final blended image. I am not sure about the feasibility; however, this would be an interesting experiment. 

Additionally, I also was intrigued to know if participants having good technical (CS background) or artistic background (Graphics) are faster in finding images. I believe that the more “internet searching” talent we have, the lesser the time. 

Questions

  1. What other fields can this idea be extended to? Could we possibly find texts related to the blended image? For example, for New-York and night, we could find articles that talk about both the concepts. This may result in finding more images too. 
  2. Would using complex images like scenes as opposed to only singular images help? Computer Vision algorithms can be used to get bounds boxes out of such images to get individual objects. 
  3. Do people with better technical or graphical skills perform better?

Read More

4/29/20 – Lee Lisle – Visiblends: A Flexible Workflow for Visual Blends

Summary

Chilton et al.’s paper describes a flexible workflow for creating visual blends that the authors dub “Visiblends’ (do you see what they did there?). Lack of imagination in naming notwithstanding, the workflow involves an input of two concepts, brainstorming, image classification and blending, and evaluation of the automatic blending. They performed three studies where their workflow was tested, showing that decentralized groups of people could appropriately brainstorm and generate blends in microtasks. The first phase involves associating different words with the input themes to create a broad base of kinds of images. The second phase involved the searching for related imagery. The third phase asked crowdworkers to annotate the found images for basic shapes and coverages of those shapes. The fourth stage is performed by an AI and involves shape matching between images to combine the two themes, while the final stage (also by AI) blends the images based on the image matching. Their studies confirm that decentralized groups, collaborative groups, and novices can all use this workflow to create visual blends.

Personal Reflection

               I liked this work, overall, as a way for people to get interesting designs out of a few keywords of whatever they’re working on. I was somewhat surprised that the second step (“Finding Images”) was not an automatic process. I had figured when I read the introduction that this step was automated by image recognition software, since these are not complex images but images of single objects. However, when it was explained in the Workflow section, it makes it clear that these images are essentially another phase of the brainstorming process. However, I was concerned that it was perhaps a complex microtask since it asked for an implementation of several somewhat complex filters as well as ten images from those filters.

               I thought the images in Figure 7 were somewhat deceptive, however. They stated in the caption for that image that there was “aesthetic editing by an artist,” which implies they had a visual designer already employed. If that was the case, why is the expert not performing the expert task? I would have liked to see the actual resultant images as they show (some of) in the later studies.

               The refinement process they introduced in the first study was also interesting in that the refinement was more than just asking for more results – the user actually iterated on the design process to find similar shaped items between the two categories. This shows an aspect of human intelligence to solve a problem that the AI had difficulty solving – realizing why the AI was having trouble was a key part of the process.

               Lastly, I would have liked to see what would have happened if all three groups (decentralized, group collaboration, and novices) were given the same concepts to generate ideas from. Which group might have performed the best? Also, since this is quite decentralized, I would have liked to see an mTurk deployment to see how the crowd could perform this task as well.

Questions

  1. As discussed above, an interesting part of this paper was how human intelligence was employed to refine the AI’s process, thereby giving it better inputs. Are there other ways that using human insight into why an AI is having issues is a good way to solve the problem?
  2. When a workflow that creates microtasks like this, is it more helpful to test it with participants that come into a lab or through a crowdworking site like Mechanical Turk? Should they both be performed? Why or why not?
  3. Would you use a service like this for a creative image in your work? Why or why not?

Read More

4/29/2020 – Akshita Jha – Accelerating Innovation Through Analogy Mining

Summary:
“Accelerating Innovation Through Analogy Mining” by Hope et al. talks about the problem of analogy mining from messy and chaotic real-world datasets. Hand created databases have high relational structures but are sparse in nature. On the other hand, machine learning and information retrieval techniques can be used but they lack the understanding of the underlying structure which is crucial to analogy related tasks. The authors leverage the strengths of both crowdsourcing and machine learning techniques to learn analogies from these real-world datasets. They make use of the creativity of the crowds with the cheap computing power of recurrent neural networks. The authors extract meaningful vector representations from product descriptions. They observe that this methodology achieves greater precision and recall than the traditional information-retrieval methods. The authors also demonstrate that the models significantly helped in generating more creative ideas compared to analogies retrieved by traditional methods.

Reflections:
This is a really interesting paper that talks about a scalable approach to finding analogies in large, real-world, messy datasets. The authors use a bi-directional Recurrent Neural Network (RNN) with Gated Recurrent Units (GRU) to learn the purpose and mechanism vectors for product descriptions. However, since the paper came out there have been great advances in the field of natural language processing tasks because of BERT: Bidirectional Encoder Representations from Transformers. BERT has achieved a state of the art results for many natural language tasks like question answering, natural language understanding, search, and retrieval, etc.. I’m curious to know how BERT would affect the results of this current system. Would we still need crowd workers for analogy detection or would using BERT alone for analogy computation suffice? One of the limitations of RNN is that it is directional, i.e., it can either read from right to left or left to right, or both. BERT is essentially non-directional, i.e, it takes all the words as inputs at once and hence, can compute the complex non-linear relationships between them. This would definitely prove helpful for detecting analogies. the approach taken by the authors by using TF-IDF did result in diversity but did not take into account the relevance. Also, the purpose and mechanism vector by the authors did not distinguish between high and low-level features. These learned vectors also did not take into account the intra-dependencies between different purposes and mechanisms or the inter-dependency between various purposes and mechanisms. It would be interesting to observe how these dependencies could be encoded and whether they would benefit the final tasks of analogy computation. Another aspect that can be looked into is the trade-off between generating useful vector and its relation with the creativity of the crowd-workers. Does creativity increase, decrease, or remain the same?

Questions:
1. Which another creative task can benefit from human-AI interaction?
2. Why is the task of analogy computation important?
3. How are you incorporating and leveraging the strength of crowd-workers and machine learning in your project?

Read More